Thread: Btree runtime recovery. Stuck spins.

Btree runtime recovery. Stuck spins.

From
"Mikheev, Vadim"
Date:
Hi

1. Subj is implemented and is ON by default in current.
There is flag - fixbtree - to turn this feature OFF.
I've run some tests: 100 clients inserted big tuples (1700-1800
bytes) into single table with index. After splitting root page
elog(ERROR) was forced, as well as after each ~ 5th non-root page
split, so there was what to fix. After this I've run selects for
each key to check that index structure is correct.

2. During tests I've got stuck spin aborts couple of times.
So I've increased S_MAX_BUSY, placed elog(NOTICE, "WOULD BE STUCK")
for spins == 20001 in s_lock_sleep() and rerun tests.
I've got *many* "WOULD BE STUCK" notices but no one abort.

Does it explain why I tried to avoid spin stuck "detection" code
in WAL? -:)

Shouldn't we increase S_MAX_BUSY and use ERROR instead of FATAL?

Vadim


Re: Btree runtime recovery. Stuck spins.

From
Tom Lane
Date:
"Mikheev, Vadim" <vmikheev@SECTORBASE.COM> writes:
> 2. During tests I've got stuck spin aborts couple of times.
> So I've increased S_MAX_BUSY, placed elog(NOTICE, "WOULD BE STUCK")
> for spins == 20001 in s_lock_sleep() and rerun tests.
> I've got *many* "WOULD BE STUCK" notices but no one abort.
> Does it explain why I tried to avoid spin stuck "detection" code
> in WAL? -:)
> Shouldn't we increase S_MAX_BUSY and use ERROR instead of FATAL?

No.  If you have delays exceeding a minute, or that are even a visible
fraction of a minute, then a spinlock is NOT the correct mechanism to be
using to wait ... because guess what, it's spinning, and consuming
processor time to no purpose.  You should be using a lock instead for
anything that involves more than a trivial amount of delay.
        regards, tom lane


Re: Btree runtime recovery. Stuck spins.

From
"Vadim Mikheev"
Date:
> > Shouldn't we increase S_MAX_BUSY and use ERROR instead of FATAL?
> 
> No. If you have delays exceeding a minute, or that are even a visible
> fraction of a minute, then a spinlock is NOT the correct mechanism to be
> using to wait ... because guess what, it's spinning, and consuming
> processor time to no purpose.  You should be using a lock instead for
> anything that involves more than a trivial amount of delay.

"Amount of delay" depends on system load - something we can't control...

Btree uses spins to lock buffers (as all other access methods) and so
I could use only spins in new code. And though tree recovery locks buffers
for longer time than normal insert operations it's possible to get
"stuck" spins when using concurrent buffers locks *everywhere* under
heavy load (especially with WAL which requires holding buffer locks
for duration of logging).

So, probably we have to use some kind of light locks (without deadlock
detection) for buffers, in future.

Vadim




Re: Btree runtime recovery. Stuck spins.

From
Tom Lane
Date:
"Vadim Mikheev" <vmikheev@sectorbase.com> writes:
> Shouldn't we increase S_MAX_BUSY and use ERROR instead of FATAL?
>> 
>> No. If you have delays exceeding a minute, or that are even a visible
>> fraction of a minute, then a spinlock is NOT the correct mechanism to be
>> using to wait ... because guess what, it's spinning, and consuming
>> processor time to no purpose.  You should be using a lock instead for
>> anything that involves more than a trivial amount of delay.

> "Amount of delay" depends on system load - something we can't control...

> Btree uses spins to lock buffers (as all other access methods) and so
> I could use only spins in new code. And though tree recovery locks buffers
> for longer time than normal insert operations it's possible to get
> "stuck" spins when using concurrent buffers locks *everywhere* under
> heavy load (especially with WAL which requires holding buffer locks
> for duration of logging).

Hm.  It was OK to use spinlocks to control buffer access when the max
delay was just the time to read or write one disk page.  But it sounds
like we've pushed the code way past what it was designed to do.  I think
this needs some careful thought, not just a quick hack like increasing
the timeout interval.
        regards, tom lane


Re: Btree runtime recovery. Stuck spins.

From
Tom Lane
Date:
I wrote:
> "Vadim Mikheev" <vmikheev@sectorbase.com> writes:
>> Btree uses spins to lock buffers (as all other access methods) and so
>> I could use only spins in new code. And though tree recovery locks buffers
>> for longer time than normal insert operations it's possible to get
>> "stuck" spins when using concurrent buffers locks *everywhere* under
>> heavy load (especially with WAL which requires holding buffer locks
>> for duration of logging).

> Hm.  It was OK to use spinlocks to control buffer access when the max
> delay was just the time to read or write one disk page.  But it sounds
> like we've pushed the code way past what it was designed to do.  I think
> this needs some careful thought, not just a quick hack like increasing
> the timeout interval.

After thinking more about this, simply increasing S_MAX_BUSY is clearly
NOT a good answer.  If you are under heavy load then processes that are
spinning are making things worse, not better, because they are sucking
CPU cycles that would be better spent on the processes that are holding
the locks.

It would not be very difficult to replace the per-disk-buffer spinlocks
with regular lockmanager locks.  Advantages: * Processes waiting for a buffer lock aren't sucking CPU cycles. *
Deadlockswill be detected and handled reasonably.  (The more stuff   that WAL does while holding a buffer lock, the
biggerthe chances   of deadlock.  I think this is a significant concern now.)
 
Of course the major disadvantage is: * the lock setup/teardown overhead is much greater than for a   spinlock, and the
overheadis just wasted when there's no contention.
 

A reasonable alternative would be to stick with the spinlock mechanism,
but use a different locking routine (maybe call it S_SLOW_LOCK) that is
designed to deal with locks that may be held for a long time.  It would
use much longer delay intervals than the regular S_LOCK code, and would
have either a longer time till ultimate timeout, or no timeout at all.
The main problem with this idea is choosing an appropriate timeout
behavior.  As I said, I am concerned about the possibility of deadlocks
in WAL-recovery scenarios, so I am not very happy with the thought of
no timeout at all.  But it's hard to see what a reasonable timeout would
be if a minute or more isn't enough in your test cases; seems to me that
that suggests that for very large indexes, you might need a *long* time.

Comments, preferences, better ideas?
        regards, tom lane


Re: Btree runtime recovery. Stuck spins.

From
Bruce Momjian
Date:
> > Hm.  It was OK to use spinlocks to control buffer access when the max
> > delay was just the time to read or write one disk page.  But it sounds
> > like we've pushed the code way past what it was designed to do.  I think
> > this needs some careful thought, not just a quick hack like increasing
> > the timeout interval.
> 
> After thinking more about this, simply increasing S_MAX_BUSY is clearly
> NOT a good answer.  If you are under heavy load then processes that are
> spinning are making things worse, not better, because they are sucking
> CPU cycles that would be better spent on the processes that are holding
> the locks.

Our spinlocks don't go into an infinite test loop, right?  They back off
and retest at random intervals.

I can't imagine we don't have similar btree lock needs other places in
the code were a solution already exists.

--  Bruce Momjian                        |  http://candle.pha.pa.us pgman@candle.pha.pa.us               |  (610)
853-3000+  If your life is a hard drive,     |  830 Blythe Avenue +  Christ can be your backup.        |  Drexel Hill,
Pennsylvania19026
 


Re: Btree runtime recovery. Stuck spins.

From
Tom Lane
Date:
Bruce Momjian <pgman@candle.pha.pa.us> writes:
> Our spinlocks don't go into an infinite test loop, right?  They back off
> and retest at random intervals.

Not very random --- either 0 or 10 milliseconds.  (I think there was
some discussion of changing that, but it died off without agreeing on
anything.)  The point though is that the code is tuned for use with
spinlocks protecting shared-memory data structures, where no one is
supposed to be holding the lock for long; so a short retry delay is
appropriate, and we don't need a very long time before declaring "stuck
spinlock" either.  This is not optimal for cases where someone may be
holding the lock for a good while, but detuning the code so that it
works less well for the shmem-structure case isn't the answer.

We need two different mechanisms.
        regards, tom lane


Re: Btree runtime recovery. Stuck spins.

From
ncm@zembu.com (Nathan Myers)
Date:
On Fri, Feb 09, 2001 at 01:23:35PM -0500, Tom Lane wrote:
> Bruce Momjian <pgman@candle.pha.pa.us> writes:
> > Our spinlocks don't go into an infinite test loop, right?  They back off
> > and retest at random intervals.
> 
> Not very random --- either 0 or 10 milliseconds.  (I think there was
> some discussion of changing that, but it died off without agreeing on
> anything.) ...

I think we agreed that 0 was just wrong, but nobody changed it.
Changing it to 1 microsecond would be the smallest reasonable 
change.  As it is, it just does a bunch of no-op syscalls each time it
wakes up after a 10ms sleep, without yielding the CPU.

Nathan Myers
ncm@zembu.com


Re: Btree runtime recovery. Stuck spins.

From
"Vadim Mikheev"
Date:
> > Hm.  It was OK to use spinlocks to control buffer access when the max
> > delay was just the time to read or write one disk page.  But it sounds

Actually, btree split requires 3 simult. buffers locks and after that
_bt_getstackbuf may read *many* parent buffers while holding locks on
2 buffers. AFAIR, the things are even worse in hash.

And anyway, there is always probability that someone else will get just
freed lock while you're waiting next 0.01 sec. The problem is that
there is no priority/ordering while waiting for spin lock.

> > like we've pushed the code way past what it was designed to do.  I think
> > this needs some careful thought, not just a quick hack like increasing
> > the timeout interval.

I fear there is not enough time -:(

> After thinking more about this, simply increasing S_MAX_BUSY is clearly
> NOT a good answer.  If you are under heavy load then processes that are
> spinning are making things worse, not better, because they are sucking
> CPU cycles that would be better spent on the processes that are holding
> the locks.
> 
> It would not be very difficult to replace the per-disk-buffer spinlocks
> with regular lockmanager locks.  Advantages:
>   * Processes waiting for a buffer lock aren't sucking CPU cycles.
>   * Deadlocks will be detected and handled reasonably.  (The more stuff
>     that WAL does while holding a buffer lock, the bigger the chances
>     of deadlock.  I think this is a significant concern now.)

I disagree. Lmgr needs in deadlock detection code because of deadlock
may be caused by *user application* design and we must not count on
*user application* correctness. But we must not use deadlock detection
code when we protected from deadlock by *our* design. Well, anyone can
make mistake and break order of lock acquiring - we should just fix
those bugs -:)
So, it doesn't matter *how much stuff that WAL does while holding buffer
locks* as long as WAL itself doesn't acquire buffer locks.

> Of course the major disadvantage is:
>   * the lock setup/teardown overhead is much greater than for a
>     spinlock, and the overhead is just wasted when there's no contention.

Exactly.

> A reasonable alternative would be to stick with the spinlock mechanism,
> but use a different locking routine (maybe call it S_SLOW_LOCK) that is
> designed to deal with locks that may be held for a long time.  It would
> use much longer delay intervals than the regular S_LOCK code, and would
> have either a longer time till ultimate timeout, or no timeout at all.
> The main problem with this idea is choosing an appropriate timeout
> behavior.  As I said, I am concerned about the possibility of deadlocks
> in WAL-recovery scenarios, so I am not very happy with the thought of
> no timeout at all.  But it's hard to see what a reasonable timeout would

And I'm unhappy with timeouts -:) It's not solution at all. We should
do right design instead.

> be if a minute or more isn't enough in your test cases; seems to me that
> that suggests that for very large indexes, you might need a *long* time.
> 
> Comments, preferences, better ideas?

For any spins which held while doing IO ops we should have queue of
waiting backend' PROCs. As I said - some kind of lightweight lock manager.
Just two kind of locks - shared & exclusive. No structures to find locked
objects. No deadlock detection code. Backends should wait on their
semaphores, without timeouts.

For "true" spins (held for really short time when accessing control
structures in shmem) we should not sleep 0.01 sec! tv_usec == 1 would be
reasonable - just to yield CPU. Actually, mutexes would be much better...

Vadim