Re: Reducing tuple overhead - Mailing list pgsql-hackers

From Robert Haas
Subject Re: Reducing tuple overhead
Date
Msg-id CA+TgmoaDjD061xe5aFoXQuONKszqkVu=R0vFsO2iy-tkwOBk-A@mail.gmail.com
Whole thread Raw
In response to Re: Reducing tuple overhead  (Amit Kapila <amit.kapila16@gmail.com>)
Responses Re: Reducing tuple overhead
List pgsql-hackers
On Thu, Apr 30, 2015 at 12:31 AM, Amit Kapila <amit.kapila16@gmail.com> wrote:
> I think I am missing something here, but when this second
> evaluation is needed.  Basically what I understand from index
> insertion is that it evaluates the value to be inserted in index
> before calling nbtree module and then nbtree just inserts the
> value/tuple passed to it.

Sure, but what happens if it doesn't evaluate to the same value?

Consider a tuple where a = 1 and a function f(a).  You insert the
tuple, evaluate f(a), and get 17.  So you insert an index tuple into
the btree with a value of 17, pointing at the tuple.  Now you delete
the tuple, evaluate f(a) again, and this time you get 42.  You search
the btree for an index tuple with a value of 42, and you don't find
one.  But the index tuple is still there.

With the current approach, that doesn't happen, because we effectively
search the entire index for tuples pointing at the heap tuple we're
trying to get rid of.  The only problem with that is that it's
crushingly expensive when the index is large and the number of tuples
we're cleaning out is comparatively small.  But that's a big problem.

-- 
Robert Haas
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company



pgsql-hackers by date:

Previous
From: Robert Haas
Date:
Subject: Re: alternative compression algorithms?
Next
From: Robert Haas
Date:
Subject: Re: CTE optimization fence on the todo list?