Thread: Free Space Map data structure

Free Space Map data structure

From
Heikki Linnakangas
Date:
The last thread about Free Space Map evolved into discussion about 
whether the FSM and other kinds of auxiliary data should be stored 
within the heap pages, in "map forks", or auxiliary relfilenodes 
attached to the relation. It seems the consensus was to go with the map 
forks, but what I really wanted to discuss is the data structure needed 
for the Free Space Map.

The FSM data structure needs to support two basic operations:

1. Fast lookup of page with >= X bytes of free space
2. Update of arbitrary, individual pages.

Our current code doesn't support 2, as we always update the FSM in bulk 
after vacuum, but we will need that capability to be able to do partial 
vacuums in the future.

Additionally, the data structure needs to be "pageable", so that we can 
efficiently store it in pages that can be loaded in the buffer cache on 
demand, and not require a fixed size shared memory block.

The simplest conceivable data structure is a straight array, where nth 
element is the free space on page n. That's easily pageable, and 
provides O(1) lookup/update of arbitrary page. Unfortunately it's slow 
to search for a page with X bytes of free space in that.

One brilliant idea I had, is a binary heap/tree kind of structure, where 
each heap page is represented by one leaf node. Each leaf node stores 
the amount of free space on the corresponding heap apge. Each non-leaf 
node stores the max. amount of free space in any of its children. So the 
top-most node immediately tells the max. amount of free space on *any* 
page, which means that to find out that there's no suitable page is a 
O(1) operation, which is good when you're inserting to a full relation. 
When you're looking for X bytes, you traverse the tree down a path with 
nodes > X.

For example:
    9 4     9
2 4   0 9

The leaf nodes correspond the heap pages, so page #0 has 2 units of free 
space, page #1 has 4, page #1 is full and page has 9.

Let's work through a couple of examples:

To search for a page with 3 bytes of free space, start from the top. We 
see that the topmost node has value 9, so we know there is a page 
somewhere with enough space. Traverse down the tree, to the node where X >= 3. In this case, that's the left child, but
ifit was true for both, 
 
we could pick either one. Traverse down from that node similarly, until 
we hit the bottom.

To update the free space on page #1 to 3, you look up the leaf node 
corresponding that page, which is easy if we store the tree in an array. 
We update the 4 on that node to 3, and walk up the tree updating the 
parents. In this case, we update the parent of that node to 3, and stop, 
because the value in top node is higher than that. The lookup/update is 
therefore O(log N) operation.


Unfortunately, this data structure isn't easily pageable. It still seems 
like a good and simple structure within a page, but we need a way to 
scale it.

If we use one byte to store the amount of free space on a page (I think 
that's enough granularity), you can fit a tree like that with about 8000 
nodes, with 4000 leaf nodes, in one 8k block. That means that we can 
address 4000 heap pages, or 32MB of heap, with one FSM page like that.

To go above that, we can introduce upper pages, where the leaf nodes 
refer not to heap pages but to other FSM pages. The addressing gets 
tricky, though. Because the number of upper pages required depends on 
the depth of the tree, we can't just divide heap page number by 4000 to 
get to the right FSM page. I think that's solvable, by always storing 
the upper pages at predetermined places in the FSM file, but you'll need 
a bit of logic to get from heap page # to FSM page #.

Thoughts? Any other data structures that better fill the bill?

--   Heikki Linnakangas  EnterpriseDB   http://www.enterprisedb.com


Re: Free Space Map data structure

From
"Brendan Jurd"
Date:
-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1

On Tue, Apr 8, 2008 at 6:33 PM, Heikki Linnakangas  wrote:
>  For example:
>
>     9
>   4     9
>  2 4   0 9
>
>  The leaf nodes correspond the heap pages, so page #0 has 2 units of free
> space, page #1 has 4, page #1 is full and page has 9.
>
>  Let's work through a couple of examples:
>
>  To search for a page with 3 bytes of free space, start from the top. We see
> that the topmost node has value 9, so we know there is a page somewhere with
> enough space.

If I understand your design correctly, this claim isn't true.  If the
topmost node reports 9 bytes free, could you not have seven pages each
with 1 byte free, and an eighth page with 2 bytes free?
         9   4          52    2    2    3
1 1 1 1 1 1 1 2

So you'd actually end up walking two levels down the left hand side of
the tree, discovering not enough space, and then walking two levels
down the right hand side to again discover not enough space.

(Apologies if the above is somehow misguided)

Otherwise seems like a cool structure.

Cheers,
BJ
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.7 (GNU/Linux)
Comment: http://getfiregpg.org

iD8DBQFH+zLI5YBsbHkuyV0RAhD+AKDiH3S2vG7bSYMR6JvmUK5nfK/5zQCaA+kA
Gc1wGRty/5zBvRqDBZ/pt+4=
=5tcB
-----END PGP SIGNATURE-----


Re: Free Space Map data structure

From
"Dave Page"
Date:
On Tue, Apr 8, 2008 at 9:54 AM, Brendan Jurd <direvus@gmail.com> wrote:
>  If I understand your design correctly, this claim isn't true.  If the
>  topmost node reports 9 bytes free, could you not have seven pages each
>  with 1 byte free, and an eighth page with 2 bytes free?
>
>           9
>     4          5
>   2    2    2    3
>  1 1 1 1 1 1 1 2
>
>  So you'd actually end up walking two levels down the left hand side of
>  the tree, discovering not enough space, and then walking two levels
>  down the right hand side to again discover not enough space.

As I read it, each node takes the value of the largest child, not the
sum of the children.


-- 
Dave Page
EnterpriseDB UK: http://www.enterprisedb.com


Re: Free Space Map data structure

From
Hannu Krosing
Date:
On Tue, 2008-04-08 at 09:33 +0100, Heikki Linnakangas wrote:
> The last thread about Free Space Map evolved into discussion about 
> whether the FSM and other kinds of auxiliary data should be stored 
> within the heap pages, in "map forks", or auxiliary relfilenodes 
> attached to the relation. It seems the consensus was to go with the map 
> forks, but what I really wanted to discuss is the data structure needed 
> for the Free Space Map.
> 
> The FSM data structure needs to support two basic operations:
> 
> 1. Fast lookup of page with >= X bytes of free space
> 2. Update of arbitrary, individual pages.
> 
> Our current code doesn't support 2, as we always update the FSM in bulk 
> after vacuum, but we will need that capability to be able to do partial 
> vacuums in the future.
> 
> Additionally, the data structure needs to be "pageable", so that we can 
> efficiently store it in pages that can be loaded in the buffer cache on 
> demand, and not require a fixed size shared memory block.
> 
> The simplest conceivable data structure is a straight array, where nth 
> element is the free space on page n. That's easily pageable, and 
> provides O(1) lookup/update of arbitrary page. Unfortunately it's slow 
> to search for a page with X bytes of free space in that.



> One brilliant idea I had, is a binary heap/tree kind of structure, where 
> each heap page is represented by one leaf node. Each leaf node stores 
> the amount of free space on the corresponding heap apge. Each non-leaf 
> node stores the max. amount of free space in any of its children. So the 
> top-most node immediately tells the max. amount of free space on *any* 
> page, which means that to find out that there's no suitable page is a 
> O(1) operation, which is good when you're inserting to a full relation. 
> When you're looking for X bytes, you traverse the tree down a path with 
> nodes > X.
> 
> For example:
> 
>      9
>   4     9
> 2 4   0 9
> 
> The leaf nodes correspond the heap pages, so page #0 has 2 units of free 
> space, page #1 has 4, page #1 is full and page has 9.
> 
> Let's work through a couple of examples:
> 
> To search for a page with 3 bytes of free space, start from the top. We 
> see that the topmost node has value 9, so we know there is a page 
> somewhere with enough space. Traverse down the tree, to the node where X 
>  >= 3. In this case, that's the left child, but if it was true for both, 
> we could pick either one. Traverse down from that node similarly, until 
> we hit the bottom.
> 
> To update the free space on page #1 to 3, you look up the leaf node 
> corresponding that page, which is easy if we store the tree in an array. 
> We update the 4 on that node to 3, and walk up the tree updating the 
> parents. In this case, we update the parent of that node to 3, and stop, 
> because the value in top node is higher than that. The lookup/update is 
> therefore O(log N) operation.
> 
> 
> Unfortunately, this data structure isn't easily pageable. It still seems 
> like a good and simple structure within a page, but we need a way to 
> scale it.

Also, we should aim at not needing write anything for "initially full"
pages, so that there will be no lock contention for parallel bulkloads.

That should be easily achieved, if we initialize all pages to "full" and
only do the tree update only if there is some room (above a certain
threshold) left after doing an insert.

In other words, pages filled by bulkload should not go into FSM.

> If we use one byte to store the amount of free space on a page (I think 
> that's enough granularity), you can fit a tree like that with about 8000 
> nodes, with 4000 leaf nodes, in one 8k block. That means that we can 
> address 4000 heap pages, or 32MB of heap, with one FSM page like that.
> 
> To go above that, we can introduce upper pages, where the leaf nodes 
> refer not to heap pages but to other FSM pages. The addressing gets 
> tricky, though. Because the number of upper pages required depends on 
> the depth of the tree, we can't just divide heap page number by 4000 to 
> get to the right FSM page. I think that's solvable, by always storing 
> the upper pages at predetermined places in the FSM file, but you'll need 
> a bit of logic to get from heap page # to FSM page #.

If we could rely on sparse files, then we could just extend the
tree-in-array idea to pages. Two levels of of pages (4000) can cover
256Gb of data ((8k*8k)/2) leaves * 8kb page size

The pages actually containing any meaningful data can be computed, the
rest should not be touched. 

Same is true for in-page updates - if we have only 4 pages in heap, then
there is no need to update the whole first branch (12 levels) just
update the page and its 2 parents, and when searching, start from parent
when adding 5th page, then extend the tree upward and start future
searches from there 3rd level grandparent

Probably we could do without sparse files, if we find an efficient way
to compute the "add order" of leaf and parent pages for above algorithm.

> Thoughts? Any other data structures that better fill the bill?



> -- 
>    Heikki Linnakangas
>    EnterpriseDB   http://www.enterprisedb.com
> 



Re: Free Space Map data structure

From
"Pavan Deolasee"
Date:
On Tue, Apr 8, 2008 at 2:03 PM, Heikki Linnakangas
<heikki@enterprisedb.com> wrote:

>  Our current code doesn't support 2, as we always update the FSM in bulk
> after vacuum, but we will need that capability to be able to do partial
> vacuums in the future.
>

+1 for that. This will be extremely useful to update FSM after HOT prune/defrag.
Right now, the page free space is only available for subsequent UPDATEs in
the page.

>
>  Thoughts? Any other data structures that better fill the bill?
>

Can we not use bitmaps to track approximate  rather than exact free
space ? For example, we can have 8 or 16 buckets of free space.
A page with 1-32 bytes free, goes into bucket 1, a page with at 33-64 bytes
free, goes into bucket 2 and so on. If you want a page with X bytes free,
look into the next immediate larger bucket. (the ranges can varry here)

This would give us O(1) time for FSM updates. To speed up lookup, we
can use hierarchical bitmaps. Lets work through an example for 8 buckets:

At the segment level, we have one FSM page. That can summarize FSM
info for ~8000 heap page (1 bit per page per bucket). In addition, we
can have first level hierarchy in the same page where one  bit, say
summarizes info for 100 pages. So if there a page with 'bucket' worth
of free space in the first 100 pages in the segment, the corresponding
bit in the first hierarchy will be set. In this case, the first level hierarchy
would need 80 bits per bucket i.e 80 bytes.

We can then extend the concept of segments to say, 'extents'. One FSM page
per extent summarizes the FSM info for segments in that extent. With the same
logic as above, we can have ~8000 segments per extent.

And then one FSM page to summarize info for all extents. I guess at that
level we would run out of maximum supported file size anyways.

Well, this could be completely orthogonal to suggestions you are seeking,
but nevertheless I had this thought for quite some time. So just wanted
to speak about it :-)

Thanks,
Pavan




-- 
Pavan Deolasee
EnterpriseDB http://www.enterprisedb.com


Re: Free Space Map data structure

From
Hannu Krosing
Date:
On Tue, 2008-04-08 at 12:26 +0300, Hannu Krosing wrote:

> Probably we could do without sparse files, if we find an efficient way
> to compute the "add order" of leaf and parent pages for above algorithm.

if we always add only the minimal needed set of parents then the order
will look something like
1: 02: 1 3: (0-1) 4: 2 5: (0-3) 6: 3 7: (2-3) 8: 49: (0-7)
10: 5
11: (4-5)
12. 6
13: (4-7)
13: 7
14: (6-7)

seems pretty regular  :)

and is probably easy to expand into pages

------------
Will work on this a little more



Re: Free Space Map data structure

From
Gregory Stark
Date:
"Heikki Linnakangas" <heikki@enterprisedb.com> writes:

> For example:
>
>     9
>  4     9
> 2 4   0 9

It occurs to me now that this it actually wouldn't be so easy to ripple up
changes due to the amount amount of space *decreasing*. Consider if you reduce
the leaf 9 to an 8. You want to decrease its parent to 8 but first you have to
check that its sibling isn't also 9. Again when you get to the grandparnt you
have to check that the "4" isn't a 9. As long as you're on one page that's
cheap but if it means paging in another page of the FSM that could be
annoying.

--  Gregory Stark EnterpriseDB          http://www.enterprisedb.com Ask me about EnterpriseDB's PostGIS support!


Re: Free Space Map data structure

From
Hannu Krosing
Date:
On Tue, 2008-04-08 at 13:38 +0300, Hannu Krosing wrote:
> On Tue, 2008-04-08 at 12:26 +0300, Hannu Krosing wrote:
> 
> > Probably we could do without sparse files, if we find an efficient way
> > to compute the "add order" of leaf and parent pages for above algorithm.
> 
> if we always add only the minimal needed set of parents then the order
> will look something like
> 
>  1: 0
>  2: 1 
>  3: (0-1) 
>  4: 2 
>  5: (0-3) 
>  6: 3 
>  7: (2-3) 
>  8: 4
>  9: (0-7)
> 10: 5
> 11: (4-5)
> 12. 6
> 13: (4-7)
> 13: 7
> 14: (6-7)
> 
> seems pretty regular  :)
> 
> and is probably easy to expand into pages
> 
> ------------
> Will work on this a little more
 index   tree node  binary mask  mask bits   0        0          00000            1        0-1        0000?         1
2       1          00001            3        0-3        000??         2   4        2          00010            5
2-3       0001?         1   6        3          00011            7        0-7        00???         3   8        4
  00100            9        4-5        0010?         1  10        5          00101           11        4-7        001??
       2  12        6          00110           13        6-7        0011?         1  14        7          00111
 15        0-15       0????         4  16        8          01000           17        8-9        0100?         1  18
   9          01001           19        8-11       010??         2  20        10         01010           21
10-11     0101?         1  22        11         01011           23        8-15       01???         3  24        12
  01100           25        12-13      0110?         1  26        13         01101           27        12-15      011??
       2  28        14         01110           29        14-15      0111?         1  30        15         01111
 31        0-31       ?????         5
 
...32........16.........10000.........

seems to be a simple pattern

for each node row (each second row) take binary representation of 
next row, start from end and replace all zeros up to and including 
rightmost one with mask bits.

mask 011?? means (data and 11100) == 01100

I will try to work out some experimental/sample code in python for find
and update in the evening.

---------------
Hannu





Re: Free Space Map data structure

From
Tom Lane
Date:
Heikki Linnakangas <heikki@enterprisedb.com> writes:
> ... what I really wanted to discuss is the data structure needed 
> for the Free Space Map.

> The FSM data structure needs to support two basic operations:
> 1. Fast lookup of page with >= X bytes of free space
> 2. Update of arbitrary, individual pages.
> Our current code doesn't support 2, as we always update the FSM in bulk 
> after vacuum, but we will need that capability to be able to do partial 
> vacuums in the future.

Note that the lack of such infrastructure is mainly because we didn't
need it, not because it couldn't be done.  But in any case the current
implementation is a loser, agreed.

> One brilliant idea I had, is a binary heap/tree kind of structure, where 
> each heap page is represented by one leaf node.

I'm worried about a couple of things here:

* Loss of concurrency in access to the FSM itself, due to updates having
to change large amounts of interrelated entries, and due to all
processes always wanting to touch the root first.

* Loss of concurrency in subsequent heap updates.  The current FSM gets
one thing right: if different backends ask for space, they will (if at
all possible) be given different heap pages to insert into, and
therefore will not suffer page-level contention while they do their
insertions.  I don't see what you'll do to ensure a comparable behavior
with this structure.

> Unfortunately, this data structure isn't easily pageable. It still seems 
> like a good and simple structure within a page, but we need a way to 
> scale it.

I suggest that you *not* scale it.  It seems possibly useful to have
this kind of structure within a single map page, but trying to go above
that will result in too much contention IMHO.
        regards, tom lane


Re: Free Space Map data structure

From
Heikki Linnakangas
Date:
Hannu Krosing wrote:
> On Tue, 2008-04-08 at 12:26 +0300, Hannu Krosing wrote:
> 
>> Probably we could do without sparse files, if we find an efficient way
>> to compute the "add order" of leaf and parent pages for above algorithm.
> 
> if we always add only the minimal needed set of parents then the order
> will look something like
> 
>  1: 0
>  2: 1 
>  3: (0-1) 
>  4: 2 
>  5: (0-3) 
>  6: 3 
>  7: (2-3) 
>  8: 4
>  9: (0-7)
> 10: 5
> 11: (4-5)
> 12. 6
> 13: (4-7)
> 13: 7
> 14: (6-7)
> 
> seems pretty regular  :)
> 
> and is probably easy to expand into pages

That's what I thought when I started thinking about this, but even after 
spending a good few hours sketching i on a whiteboard, I still couldn't 
figure out what the actual formula behind that is. I feel like I'm 
missing something obvious textbook algorithm here, so please help me out 
if you can spot a pattern in that. If there isn't one, we could build a 
lookup table for enough levels by hand, but...

--   Heikki Linnakangas  EnterpriseDB   http://www.enterprisedb.com


Re: Free Space Map data structure

From
Heikki Linnakangas
Date:
Pavan Deolasee wrote:
> Can we not use bitmaps to track approximate  rather than exact free
> space ? For example, we can have 8 or 16 buckets of free space.
> A page with 1-32 bytes free, goes into bucket 1, a page with at 33-64 bytes
> free, goes into bucket 2 and so on. If you want a page with X bytes free,
> look into the next immediate larger bucket. (the ranges can varry here)

Yep, that's actually what I was planning to do, except I was thinking of 
using 8 bits per page, or 256 buckets, because that makes the code a 
little bit simpler. 16 buckets would probably be enough in practice, though.

Also, the free space doesn't necessarily need to be divided linearly 
into buckets, we could for example use 8 buckets like:

0    32
1    64
2    128
3    256
4    512
5    1024
6    2048
7    4096

> This would give us O(1) time for FSM updates. To speed up lookup, we
> can use hierarchical bitmaps. Lets work through an example for 8 buckets:
> 
> At the segment level, we have one FSM page. That can summarize FSM
> info for ~8000 heap page (1 bit per page per bucket). In addition, we
> can have first level hierarchy in the same page where one  bit, say
> summarizes info for 100 pages. So if there a page with 'bucket' worth
> of free space in the first 100 pages in the segment, the corresponding
> bit in the first hierarchy will be set. In this case, the first level hierarchy
> would need 80 bits per bucket i.e 80 bytes.
> 
> We can then extend the concept of segments to say, 'extents'. One FSM page
> per extent summarizes the FSM info for segments in that extent. With the same
> logic as above, we can have ~8000 segments per extent.
> 
> And then one FSM page to summarize info for all extents. I guess at that
> level we would run out of maximum supported file size anyways.

Well, if you add the higher levels, we're no longer talking about O(1), 
but O(log n) (for a pretty steep logarithm), just like my proposal.

> Well, this could be completely orthogonal to suggestions you are seeking,
> but nevertheless I had this thought for quite some time.

It's actually not that orthogonal. You describe it as a hierarchical 
bitmap, but it's essentially the same idea as the binary heap/tree, just 
with way more than 2 children per parent.

--   Heikki Linnakangas  EnterpriseDB   http://www.enterprisedb.com


Re: Free Space Map data structure

From
"Pavan Deolasee"
Date:
On Wed, Apr 9, 2008 at 12:22 AM, Heikki Linnakangas
<heikki@enterprisedb.com> wrote:

>
>  Well, if you add the higher levels, we're no longer talking about O(1), but
> O(log n) (for a pretty steep logarithm), just like my proposal.
>

For updates, I would still call it O(1) because the number of levels is limited
and a very small number.


>
>  It's actually not that orthogonal. You describe it as a hierarchical
> bitmap, but it's essentially the same idea as the binary heap/tree, just
> with way more than 2 children per parent.
>

Yes, I agree.

Btw, I agree with Tom's point about the lock contention on the higher levels
for FSM updates. What we can do is during normal operations, FSM pages
are updated with a SHARE lock - so the information can best be considered
only as hint. Hopefully, if we are only doing bit flipping, we won't go wrong
often. And periodically, VACUUM would correct any mistakes in FSM info.


Thanks,
Pavan


-- 
Pavan Deolasee
EnterpriseDB http://www.enterprisedb.com


Re: Free Space Map data structure

From
Hannu Krosing
Date:
On Wed, 2008-04-09 at 10:17 +0530, Pavan Deolasee wrote:
> On Wed, Apr 9, 2008 at 12:22 AM, Heikki Linnakangas
> <heikki@enterprisedb.com> wrote:
> 
> >
> >  Well, if you add the higher levels, we're no longer talking about O(1), but
> > O(log n) (for a pretty steep logarithm), just like my proposal.
> >
> 
> For updates, I would still call it O(1) because the number of levels is limited
> and a very small number.
> 
> 
> >
> >  It's actually not that orthogonal. You describe it as a hierarchical
> > bitmap, but it's essentially the same idea as the binary heap/tree, just
> > with way more than 2 children per parent.
> >
> 
> Yes, I agree.
> 
> Btw, I agree with Tom's point about the lock contention on the higher levels
> for FSM updates. What we can do is during normal operations, FSM pages
> are updated with a SHARE lock - so the information can best be considered
> only as hint. Hopefully, if we are only doing bit flipping, we won't go wrong
> often. And periodically, VACUUM would correct any mistakes in FSM info.

Sabing 1 byte is an atomic op so if we update going from branches to
root and read from root to branches, then we can probably detect
inconsistencies when reading and then re-read.

What I'm more concerned about is L1 cache swapping between several
processors/cores, which could lead to content switch storms we used to
have in locking.

anyway, one thing to do for sure is to avoid storing info for new/last
page(s) appended by any process before that process is done filling it,
as that would lead to excessive up-tree max_free updates.

BTW, I'm pretty sure I have figured out the method for storing minimal
required binary tree as an array, where adding each page adds exactly
one upper node. The node added is often not the immediate parent, but is
always the one needed for covering all nodes.

I just have to write it up in an understandable way and then you all can
look at it and tell if it is something well-known from Knuth or Date ;)


--------------
Hannu




Re: Free Space Map data structure

From
Heikki Linnakangas
Date:
Hannu Krosing wrote:
> Sabing 1 byte is an atomic op

Unfortunately, it's not. Most if not all modern CPUs will perform one 
byte modification as "load word" + "modify word" + "save word".

--   Heikki Linnakangas  EnterpriseDB   http://www.enterprisedb.com


Re: Free Space Map data structure

From
Heikki Linnakangas
Date:
Tom Lane wrote:
> Heikki Linnakangas <heikki@enterprisedb.com> writes:
>> ... what I really wanted to discuss is the data structure needed 
>> for the Free Space Map.
> 
>> The FSM data structure needs to support two basic operations:
>> 1. Fast lookup of page with >= X bytes of free space
>> 2. Update of arbitrary, individual pages.
>> Our current code doesn't support 2, as we always update the FSM in bulk 
>> after vacuum, but we will need that capability to be able to do partial 
>> vacuums in the future.
> 
> Note that the lack of such infrastructure is mainly because we didn't
> need it, not because it couldn't be done.  

Yep. I actually remember seeing a function in freespace.c in CVS history 
to do partial updates, but it was removed at some point because it 
wasn't needed.

>> One brilliant idea I had, is a binary heap/tree kind of structure, where 
>> each heap page is represented by one leaf node.
> 
> I'm worried about a couple of things here:
> 
> * Loss of concurrency in access to the FSM itself, due to updates having
> to change large amounts of interrelated entries, and due to all
> processes always wanting to touch the root first.

When searching for free space, we start from the top, and go down from 
there. We only need to take a shared lock on the (page that contains) 
the upper nodes, so that shouldn't become a bottleneck.

When we update a node, we can stop propagating the change upwards as 
soon as we hit a node where the other child has less space than the 
current node. For example:
         5    5         3 4    5     3
4 3  2 5   2 3

To update the first leaf node from 4 to 2, we need to update it's parent 
to 3. But we don't need to update it's parent (5), and we can stop at 
that point without accessing the root.

We probably want to mark new pages as full, as Hannu suggested, to avoid 
repeatedly updating the FSM during bulk loads.

Given that the current FSM is protected by a single lock, shared by all 
relations, and always taken in exclusive mode, I'm not too worried. I'm 
not sure how long an FSM page needs to be locked in my scheme compared 
to the current FSM, but at least there will be a separate FSM for each 
relation. I do want to beat the current FSM in terms of scalability, 
though, I think I've seen FreeSpaceLock become contented on some tests.

As a little optimization, we could optimistically start from somewhere 
else than the top node when searching for free space. Perhaps from the 
top of the FSM page that contained our previous rd_targetblock. I don't 
think we need to, though.

> * Loss of concurrency in subsequent heap updates.  The current FSM gets
> one thing right: if different backends ask for space, they will (if at
> all possible) be given different heap pages to insert into, and
> therefore will not suffer page-level contention while they do their
> insertions.  I don't see what you'll do to ensure a comparable behavior
> with this structure.

Whenever we have a choice to traverse to either left or right child 
because there's heap pages with enough free space on both sides, we can 
randomize that decision to achieve the spreading.

This actually brings up another nice property this data structure has: 
instead of randomizing, we can choose the child node that's "closer" to 
some specific heap page. That can be used to keep an updated tuple 
version close to the old version, or to maintain cluster order. That's a 
goal that's at least somewhat at odds with the goal of spreading the 
inserts, of course, but the data structure has the flexibility, which is 
nice.

--   Heikki Linnakangas  EnterpriseDB   http://www.enterprisedb.com


Re: Free Space Map data structure

From
Hannu Krosing
Date:
On Wed, 2008-04-09 at 21:09 +0300, Heikki Linnakangas wrote:
> Hannu Krosing wrote:
> > Saving 1 byte is an atomic op
> 
> Unfortunately, it's not. Most if not all modern CPUs will perform one 
> byte modification as "load word" + "modify word" + "save word".

Hmm, maybe we I should change my design to modify page free info and its
parent together ? 

or what is word ? 2 bytes ? 4 bytes ? even 8 bytes ?

------------
Hannu




Re: Free Space Map data structure

From
PFC
Date:
About the FSM :
Would it be possible to add a flag marking pages where all tuples are  
visible to all transactions ? (kinda like frozen I think)This could be useful to implement index-only scans, for
count(),or to  
 
quickly skip rows when OFFSET is used, or to use only the index when the  
selected columns are all in the index. Of course if the page is flagged as  
"may contain updated tuples", then it would have to look in the heap. But,  
for tables that are not randomly updated (for instance tables that are  
mostly appended to, like forum posts, or logs, or the huge archive table,  
etc) it could save a lot of heap lookups and IO.


Re: Free Space Map data structure

From
Heikki Linnakangas
Date:
PFC wrote:
> 
>     About the FSM :
> 
>     Would it be possible to add a flag marking pages where all tuples 
> are visible to all transactions ? (kinda like frozen I think)

Ah, the visibility map. That's another line of discussion. The current 
plan is to not tie that to the FSM, but implement it separately. There's 
some infrastructure changes that are needed for both, like the "map 
forks" (see recent FSM discussions), which is why we need to have a 
design for FSM as well before we start implementing the visibility map.

It's definitely something I want to do for 8.4. Here's my rough plan:

1. Common "map forks" support
2. Rewrite FSM
3. Implement visibility map, to allow partial vacuums
4. Implement index-only scans, using the visibility map.

We'll see how far I get in the 8.4 cycle. Any help is welcome, of course 
:-).

--   Heikki Linnakangas  EnterpriseDB   http://www.enterprisedb.com


Re: Free Space Map data structure

From
PFC
Date:
> PFC wrote:
>>      About the FSM :
>>      Would it be possible to add a flag marking pages where all tuples  
>> are visible to all transactions ? (kinda like frozen I think)
>
> Ah, the visibility map. That's another line of discussion. The current  
> plan is to not tie that to the FSM, but implement it separately. There's  
> some infrastructure changes that are needed for both, like the "map  
> forks" (see recent FSM discussions), which is why we need to have a  
> design for FSM as well before we start implementing the visibility map.
>
> It's definitely something I want to do for 8.4.
Ahhhhhh, yes, yes, yes ;) yes !!!!!

>  Here's my rough plan:
> 1. Common "map forks" support
> 2. Rewrite FSM
> 3. Implement visibility map, to allow partial vacuums
> 4. Implement index-only scans, using the visibility map.
Throwing another idea that is related to partial vacuums (perhaps ?):Consider a table that is often inserted (archive,
forumposts, log,  
 
whatever), we want to CLUSTER it.In that case it would be beneficial to only cluster the "tail" of the  
table, where all the recently inserted rows are.
Example :- Table is clustered.- Insert rows in random order, update some, delete some, etc, supposing  
inserts happen at the end- Table now looks like "head":[clustered part with some holes] plus  
"tail":[rows in random order]- As long as the "tail" fits in disk cache, the random order isn't a  
problem.- So, when the "tail" reaches a certain size :- Grab it, sort it by cluster order, write it again in the heap-
Updatethe indexes in a manner similar to VACUUM (ie. bulk update)- Table now looks like "head":[clustered part with
someholes] plus  
 
"tail":[clustered]This does not remove the holes in the "head", but is this really a  
problem ? In this usage scenario, I don't think so. Regular CLUSTER could  
also be run, much less frequently than before, and it will also be much  
faster since the rows are approximately in-order already.
This approach is complimentary to the "auto-cluster" approach where the  
index is asked "where should I insert that row ?" (this will be used to  
fill the holes). Auto-cluster will work well in tables that are updated  
very often. But starting from an empty table, or an already clustered  
table, or in a mostly-insert scenario, the index will have no idea where  
to put that row...
The goodness of this approach is that- As long as the "tail" fits in RAM, sorting it is going to be very fast  
(unlike the current CLUSTER).- Bulk index updates will also be fast as long as the list of changes to  
apply to the index fits in memory.- Therefore it will block the table for much less time than good old  
CLUSTER.- Therefore it will get more use ;)
How to make it non-locking ?- Doing something like this in pseudo SQL :INSERT INTO table SELECT * FROM (DELETE FROM
tableWHERE date > last time  
 
we did this RETURNING *) ORDER BY cluster_columns;VACUUM;
That is, take the "tail" of the table (as above), sort it, insert it back  
in big chunks, and mark the old rows as deleted just like a regular delete  
would have done. Then VACUUM.
In this case you now have :"head":[clustered part with some holes] + "big hole" + "tail":[clustered  
rows]
Is the "big hole" a problem ? Probably not, it will be marked as free  
space by VACUUM and used for new inserts. A week later we get this :"head":[clustered part with some holes] + [rows in
randomorder]  
 
+ "tail":[clustered rows]Repeating the process above will make the "tail" grow and the "hole" will  
stay more or less in the same place.
Another way to do it is to use partitions :- "archive" table- "current" table
Periodically the rows from "current" are transferred to the "archive"  
table and sorted in the process. Then "current" is truncated.This works, but it is blocking, and you have the overhead
from 
 
partitioning...



















Re: Free Space Map data structure

From
Heikki Linnakangas
Date:
Hannu Krosing wrote:
> On Wed, 2008-04-09 at 21:09 +0300, Heikki Linnakangas wrote:
>> Hannu Krosing wrote:
>>> Saving 1 byte is an atomic op
>> Unfortunately, it's not. Most if not all modern CPUs will perform one 
>> byte modification as "load word" + "modify word" + "save word".
> 
> Hmm, maybe we I should change my design to modify page free info and its
> parent together ? 
> 
> or what is word ? 2 bytes ? 4 bytes ? even 8 bytes ?

It depends on architecture, I believe it's 4 bytes on 32-bit 
architectures and 8 bytes on 64-bit ones, typically. But we have to work 
on all of them.

--   Heikki Linnakangas  EnterpriseDB   http://www.enterprisedb.com


Re: Free Space Map data structure

From
Hannu Krosing
Date:
> BTW, I'm pretty sure I have figured out the method for storing minimal
> required binary tree as an array, where adding each page adds exactly
> one upper node. The node added is often not the immediate parent, but is
> always the one needed for covering all nodes.
>
> I just have to write it up in an understandable way and then you all can
> look at it and tell if it is something well-known from Knuth or Date ;)

Find sample code attached:

not optimal at all, meant as proof-of-concept.

just run the file to see how it works, some comments in the beginning.

currently it interleaves leaf and internal nodes, but it may be better
for some uses (like seqscanning leaf nodes when searching for clustered
pos) to separate them, for example having 1st 4k for lef nodes and 2nd
for internal nodes.

also, I think that having a fan-out factor above 2 (making it more like
b-tree instead of binary tree) would be more efficient due to CPU
caches, but it takes some more work to figure it out.

Of course unless someone recognizes the algorithm and can point to
textbook example ;)

---------------
Hannu


Attachment

Re: Free Space Map data structure

From
Heikki Linnakangas
Date:
Hannu Krosing wrote:
>> BTW, I'm pretty sure I have figured out the method for storing minimal
>> required binary tree as an array, where adding each page adds exactly
>> one upper node. The node added is often not the immediate parent, but is
>> always the one needed for covering all nodes.
>>
>> I just have to write it up in an understandable way and then you all can
>> look at it and tell if it is something well-known from Knuth or Date ;)
>
> Find sample code attached:
>
> not optimal at all, meant as proof-of-concept.
>
> just run the file to see how it works, some comments in the beginning.
>
> currently it interleaves leaf and internal nodes, but it may be better
> for some uses (like seqscanning leaf nodes when searching for clustered
> pos) to separate them, for example having 1st 4k for lef nodes and 2nd
> for internal nodes.
>
> also, I think that having a fan-out factor above 2 (making it more like
> b-tree instead of binary tree) would be more efficient due to CPU
> caches, but it takes some more work to figure it out.

At least it would be more storage efficient, as you wouldn't need as
many non-leaf nodes. You would need more comparisons when traversing or
updating the tree, but as you pointed out that might be very cheap
because of cache effects.

Within a page, the traditional array representation, where root is at
position 0, it's children are at 1, 2 and so forth, is actually OK. When
the depth of the tree changes, you'll need to memcpy data around and
rebuild parent nodes, but that's acceptable. Or we can fix the depth of
the tree to the max. that fits on a page, and fill it with zeros when
it's not full; all but the rightmost pages are always full anyway.

Scaling into multiple pages, we don't want to move nodes across page
boundaries when extending the relation, so we do need something like
your scheme for that. Taking the upper nodes into account, one FSM page
can hold free space information of ~4k heap (or lower level FSM) pages.
IOW, we have a fan out of 4k across pages. With a fanout like that, we
only need 3-4 levels to address over 500 TB of data.

I've attached a diagram illustrating what I have in mind. In the
diagram, each page can hold only 7 nodes, but in reality that would be ~
BLCKSZ/2, or the 4k with default block size. The heap consists of 10
pages, the bottom leaf nodes correspond the heap pages. The leaf nodes
of the upper FSM page store the max the lower FSM pages; they should
match the top nodes of the lower FSM pages. The rightmost nodes in the
tree, colored grey, are unused.

I'm pretty happy with this structure. The concurrency seems reasonably
good, though will need to figure out how exactly the locking should work.

--
   Heikki Linnakangas
   EnterpriseDB   http://www.enterprisedb.com

Attachment

Re: Free Space Map data structure

From
Heikki Linnakangas
Date:
Hannu Krosing wrote:
>   index   tree node  binary mask  mask bits
>     0        0          00000         
>     1        0-1        0000?         1
>     2        1          00001         
>     3        0-3        000??         2
>     4        2          00010         
>     5        2-3        0001?         1
>    ...
>    31        0-31       ?????         5
> ...32........16.........10000.........
> 
> seems to be a simple pattern

Oh, I see. I was thinking that the tree would be of the same height on 
all branches, but your method seems simpler. Though it meens that 
existing nodes need to be assigned to new parents as the tree grows.

Now how to extend this to n-ary trees? Assuming we use a normal binary 
tree stored in an array the usual way, within page, we can treat the FSM 
pages as nodes with fanout of (BLCKSZ - size of headers)/2.

Thanks to the large fanout, the height of that tree is not large, max 3 
levels with default block size (*) and not much more than that with 
smaller block sizes. One idea is to use a separate "fork" for each 
level, that would keep the math simple.

(*) We need to cover max 2^32 heap pages == 2^32 leaf nodes. You can fit 
~4000 leaf nodes on a page. 4000^3 > 2^32.

--   Heikki Linnakangas  EnterpriseDB   http://www.enterprisedb.com