Thread: Double sorting split patch
Hackers,
I've got my patch with double sorting picksplit impementation for GiST into more acceptable form. A little of testing is below. Index creation time is slightly higher, but search is much faster. The testing datasets were following:
1) uniform dataset - 10M rows
2) geonames points - 7.6M rows
test=# create index uniform_new_linear_idx on uniform using gist (point);
CREATE INDEX
Time: 397362,915 ms
test=# explain (analyze, buffers) select * from uniform where point <@ box(point(0.5,0.5),point(0.501,0.501));
QUERY PLAN
-------------------------------------------------------------------------------------------------------------------------------------
Bitmap Heap Scan on uniform (cost=433.27..25873.19 rows=10000 width=16) (actual time=1.407..1.448
rows=8 loops=1)
Recheck Cond: (point <@ '(0.501,0.501),(0.5,0.5)'::box)
Buffers: shared hit=39
-> Bitmap Index Scan on uniform_new_linear_idx (cost=0.00..430.77 rows=10000 width=0) (actual time=1.388..1.388 rows=8 loops=1)
Index Cond: (point <@ '(0.501,0.501),(0.5,0.5)'::box)
Buffers: shared hit=31
Total runtime: 1.527 ms
(7 rows)
test=# explain (analyze, buffers) select * from uniform where point <@ box(point(0.3,0.4),point(0.301,0.401));
QUERY PLAN
--------------------------------------------------------------------------------------------------------------------------------------
Bitmap Heap Scan on uniform (cost=433.27..25873.19 rows=10000 width=16) (actual time=0.715..0.795
rows=15 loops=1)
Recheck Cond: (point <@ '(0.301,0.401),(0.3,0.4)'::box)
Buffers: shared hit=30
-> Bitmap Index Scan on uniform_new_linear_idx (cost=0.00..430.77 rows=10000 width=0) (actual time=0.695..0.695 rows=15 loops=1)
Index Cond: (point <@ '(0.301,0.401),(0.3,0.4)'::box)
Buffers: shared hit=15
Total runtime: 0.892 ms
(7 rows)
test=# create index uniform_double_sorting_idx on uniform using gist (point);
CREATE INDEX
Time: 492796,671 ms
test=# explain (analyze, buffers) select * from uniform where point <@ box(point(0.5,0.5),point(0.501,0.501));
QUERY PLAN
-----------------------------------------------------------------------------------------------------------------------------------------
Bitmap Heap Scan on uniform (cost=445.39..25885.31 rows=10000 width=16) (actual time=0.376..0.417
rows=8 loops=1)
Recheck Cond: (point <@ '(0.501,0.501),(0.5,0.5)'::box)
Buffers: shared hit=15
-> Bitmap Index Scan on uniform_double_sorting_idx (cost=0.00..442.89 rows=10000 width=0) (actual time=0.357..0.357 rows=8 loops=1)
Index Cond: (point <@ '(0.501,0.501),(0.5,0.5)'::box)
Buffers: shared hit=7
Total runtime: 0.490 ms
(7 rows)
test=# explain (analyze, buffers) select * from uniform where point <@ box(point(0.3,0.4),point(0.301,0.401));
QUERY PLAN
------------------------------------------------------------------------------------------------------------------------------------------
Bitmap Heap Scan on uniform (cost=445.39..25885.31 rows=10000 width=16) (actual time=0.189..0.270
rows=15 loops=1)
Recheck Cond: (point <@ '(0.301,0.401),(0.3,0.4)'::box)
Buffers: shared hit=19
-> Bitmap Index Scan on uniform_double_sorting_idx (cost=0.00..442.89 rows=10000 width=0) (actual time=0.168..0.168 rows=15 loops=1)
Index Cond: (point <@ '(0.301,0.401),(0.3,0.4)'::box)
Buffers: shared hit=4
Total runtime: 0.358 ms
(7 rows)
test=# create index geonames_new_linear_idx on geonames using gist (point);
CREATE INDEX
Time: 279922,518 ms
test=# explain (analyze, buffers) select * from geonames where point <@ box(point(34.4671,126.631),point(34.5023,126.667));
QUERY PLAN
--------------------------------------------------------------------------------------------------------------------------------------
Bitmap Heap Scan on geonames (cost=341.98..19686.88 rows=7604 width=16) (actual time=0.905..0.948
rows=11 loops=1)
Recheck Cond: (point <@ '(34.5023,126.667),(34.4671,126.631)'::box)
Buffers: shared hit=25
-> Bitmap Index Scan on geonames_new_linear_idx (cost=0.00..340.07 rows=7604 width=0) (actual time=0.889..0.889 rows=11 loops=1)
Index Cond: (point <@ '(34.5023,126.667),(34.4671,126.631)'::box)
Buffers: shared hit=20
Total runtime: 1.029 ms
(7 rows)
test=# explain (analyze, buffers) select * from geonames where point <@ box(point(46.1384,-104.72), point(46.2088,-104.65));
QUERY PLAN
--------------------------------------------------------------------------------------------------------------------------------------
Bitmap Heap Scan on geonames (cost=341.98..19686.88 rows=7604 width=16) (actual time=0.644..0.776
rows=10 loops=1)
Recheck Cond: (point <@ '(46.2088,-104.65),(46.1384,-104.72)'::box)
Buffers: shared hit=13 read=6
-> Bitmap Index Scan on geonames_new_linear_idx (cost=0.00..340.07 rows=7604 width=0) (actual time=0.595..0.595 rows=10 loops=1)
Index Cond: (point <@ '(46.2088,-104.65),(46.1384,-104.72)'::box)
Buffers: shared hit=13
Total runtime: 0.857 ms
(7 rows)
test=# create index geonames_double_sorting_idx on geonames using gist (point);
CREATE INDEX
Time: 294580,774 ms
test=# explain (analyze, buffers) select * from geonames where point <@ box(point(34.4671,126.631),point(34.5023,126.667));
QUERY PLAN
------------------------------------------------------------------------------------------------------------------------------------------
Bitmap Heap Scan on geonames (cost=346.01..19690.92 rows=7604 width=16) (actual time=0.240..0.282
rows=11 loops=1)
Recheck Cond: (point <@ '(34.5023,126.667),(34.4671,126.631)'::box)
Buffers: shared hit=11
-> Bitmap Index Scan on geonames_double_sorting_idx (cost=0.00..344.11 rows=7604 width=0) (actual time=0.209..0.209 rows=11 loops=1)
Index Cond: (point <@ '(34.5023,126.667),(34.4671,126.631)'::box)
Buffers: shared hit=6
Total runtime: 0.372 ms
(7 rows)
test=# explain (analyze, buffers) select * from geonames where point <@ box(point(46.1384,-104.72), point(46.2088,-104.65));
QUERY PLAN
------------------------------------------------------------------------------------------------------------------------------------------
Bitmap Heap Scan on geonames (cost=346.01..19690.92 rows=7604 width=16) (actual time=0.311..0.352
rows=10 loops=1)
Recheck Cond: (point <@ '(46.2088,-104.65),(46.1384,-104.72)'::box)
Buffers: shared hit=13
-> Bitmap Index Scan on geonames_double_sorting_idx (cost=0.00..344.11 rows=7604 width=0) (actual time=0.293..0.293 rows=10 loops=1)
Index Cond: (point <@ '(46.2088,-104.65),(46.1384,-104.72)'::box)
Buffers: shared hit=7
Total runtime: 0.429 ms
(7 rows)
------
With best regards,
Alexander Korotkov.
Attachment
On 11.09.2011 22:30, Alexander Korotkov wrote: > Hackers, > > I've got my patch with double sorting picksplit impementation for GiST into > more acceptable form. A little of testing is below. Index creation time is > slightly higher, but search is much faster. The testing datasets were > following: > 1) uniform dataset - 10M rows > 2) geonames points - 7.6M rows I've looked at the patch, and took a brief look at the paper - but I still don't understand the algorithm. I just can't get my head around the concepts of split pairs and left/right groups. Can you explain that in layman's terms? Perhaps an example would help? -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
On Thu, Sep 15, 2011 at 7:27 PM, Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote:
------
With best regards,
Alexander Korotkov.
I've looked at the patch, and took a brief look at the paper - but I still don't understand the algorithm. I just can't get my head around the concepts of split pairs and left/right groups. Can you explain that in layman's terms? Perhaps an example would help?
In short algorithm works as following:
1) Each box can be projected to the axis as an interval. Box (x1,y1)-(x2,y2) are projected to X axis as (x1,x2) interval and to the Y axis as (y1,y2) interval. At the first step we search for splits of those intervals and select the best one.
2) At the second step produced split are converting into terms of boxes and ambiguities are solving.
2) At the second step produced split are converting into terms of boxes and ambiguities are solving.
Let's see a little deeper how intervals split search are performed by considering an example. We've intervals (0,1), (1,3), (2,3), (2,4). We assume intervals of the groups to be (0,a), (b,4). So, "a" can be some upper bound of interval: {1,3,4}, and "b" can be some lower bound of inteval: {0,1,2}.
We consider following splits: each "a" with greatest possible "b"
(0,1), (1,4)
(0,3), (2,4)
(0,4), (2,4)
and each "b" with least possible "a". In this example splits will be:
(0,1), (0,4)
(0,1), (1,4)
(0,3), (2,4)
By removing the duplicates we've following splits:
(0,1), (0,4)
(0,1), (1,4)
(0,3), (2,4)
(0,4), (2,4)
Proposed algorithm finds following splits by single pass on two arrays: one sorted by lower bound of interval and another sorted by upper bound of interval.
With best regards,
Alexander Korotkov.
On 15.09.2011 21:46, Alexander Korotkov wrote: > On Thu, Sep 15, 2011 at 7:27 PM, Heikki Linnakangas< > heikki.linnakangas@enterprisedb.com> wrote: > >> I've looked at the patch, and took a brief look at the paper - but I still >> don't understand the algorithm. I just can't get my head around the concepts >> of split pairs and left/right groups. Can you explain that in layman's >> terms? Perhaps an example would help? > > In short algorithm works as following: > 1) Each box can be projected to the axis as an interval. Box (x1,y1)-(x2,y2) > are projected to X axis as (x1,x2) interval and to the Y axis as (y1,y2) > interval. At the first step we search for splits of those intervals and > select the best one. > 2) At the second step produced split are converting into terms of boxes > and ambiguities are solving. > > Let's see a little deeper how intervals split search are performed by > considering an example. We've intervals (0,1), (1,3), (2,3), (2,4). We > assume intervals of the groups to be (0,a), (b,4). So, "a" can be some upper > bound of interval: {1,3,4}, and "b" can be some lower bound of inteval: > {0,1,2}. > We consider following splits: each "a" with greatest possible "b" > (0,1), (1,4) > (0,3), (2,4) > (0,4), (2,4) > and each "b" with least possible "a". In this example splits will be: > (0,1), (0,4) > (0,1), (1,4) > (0,3), (2,4) > By removing the duplicates we've following splits: > (0,1), (0,4) > (0,1), (1,4) > (0,3), (2,4) > (0,4), (2,4) Ok, thanks, I understand that now. > Proposed algorithm finds following splits by single pass on two arrays: one > sorted by lower bound of interval and another sorted by upper bound of > interval. That looks awfully complicated. I don't understand how that works. I wonder if two passes would be simpler? -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
On Fri, Sep 16, 2011 at 3:07 PM, Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote:
------
With best regards,
Alexander Korotkov.
That looks awfully complicated. I don't understand how that works. I wonder if two passes would be simpler?
I doubt it becomes much simpler, but here it is.
With best regards,
Alexander Korotkov.
Attachment
On 17.09.2011 17:36, Alexander Korotkov wrote: > On Fri, Sep 16, 2011 at 3:07 PM, Heikki Linnakangas< > heikki.linnakangas@enterprisedb.com> wrote: > >> That looks awfully complicated. I don't understand how that works. I wonder >> if two passes would be simpler? >> > I doubt it becomes much simpler, but here it is. Thanks! It does look a lot simpler that way, IMHO. I presume this didn't change the performance characteristics? -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
On Sat, Sep 17, 2011 at 9:00 PM, Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote:
Alexander Korotkov.
Thanks! It does look a lot simpler that way, IMHO. I presume this didn't change the performance characteristics?
On the tests I provided in the first letter difference seems to be insignificant.
------
With best regards,Alexander Korotkov.
On 15 September 2011 19:46, Alexander Korotkov <aekorotkov@gmail.com> wrote: > Proposed algorithm finds following splits by single pass on two arrays: one > sorted by lower bound of interval and another sorted by upper bound of > interval. Have you considered if further performance improvements could come from providing a qsort implementation that allows for the inlining of the comparator? I find it curious that we go to the trouble of providing a custom qsort implementation in qsort.c, pg_qsort, but haven't gone one step further and considered inlining effects. Here's a macro-based inlining implementation: http://www.corpit.ru/mjt/qsort.html I wondered in passing if compiler vendors had got around to figuring out a way of solving the "inlining function pointer" problem that I first read about years ago, so I ran this code, which benchmarks the macro-based qsort above: http://www.lemoda.net/c/inline-qsort-example/index.html It's actually C++, but that isn't significant (c stdlib qsort() is called). When I built this code with GCC 4.5, the results were: C quick-sort time elapsed: 3.24 Inline C quick-sort time elapsed: 2.01 It looks like this is something that could be revisited. -- Peter Geoghegan http://www.2ndQuadrant.com/ PostgreSQL Development, 24x7 Support, Training and Services
On Sat, Sep 17, 2011 at 9:09 PM, Peter Geoghegan <peter@2ndquadrant.com> wrote: > I find it curious that we go to the trouble of > providing a custom qsort implementation in qsort.c, pg_qsort, but > haven't gone one step further and considered inlining effects. I think we provided the qsort implementation for the benefit of platforms that didn't have a decent one and then ended up switching to use it always for some reason I don't recall. It wasn't because we thought we were better at writing qsort implementations than OS vendors. > I wondered in passing if compiler vendors had got around to figuring > out a way of solving the "inlining function pointer" problem that I > first read about years ago, so I ran this code, which benchmarks the > macro-based qsort above: You might need -fipa-pta or some other option. Or maybe LLVM would have a better chance of pulling this off. Compilers are usually pretty loath to generate specializations for every call site for fear of bloating the code. In any case I don't see how you're going to inline random database functions. Those are the cases where large amounts of data are being sorted. It's possible we sort small sets of data for internal reasons very frequently but I don't recall it turning up at the top of the profiles posted in recent days. Now a JIT that actually inlined random database functions into qsort and optimized the result would be pretty cool. But it doesn't seem like it's going to happen tomorrow... -- greg
On 18 September 2011 01:51, Greg Stark <stark@mit.edu> wrote: > I think we provided the qsort implementation for the benefit of > platforms that didn't have a decent one and then ended up switching to > use it always for some reason I don't recall. It wasn't because we > thought we were better at writing qsort implementations than OS > vendors. The comment: * We have modified their original by adding a check for already-sorted input,* which seems to be a win per discussions onpgsql-hackers around 2006-03-21. sort of suggests that it *was* at least in part because we thought we could do a better job, but that isn't a significant detail. >> I wondered in passing if compiler vendors had got around to figuring >> out a way of solving the "inlining function pointer" problem that I >> first read about years ago, so I ran this code, which benchmarks the >> macro-based qsort above: > > You might need -fipa-pta or some other option. Or maybe LLVM would > have a better chance of pulling this off. Compilers are usually pretty > loath to generate specializations for every call site for fear of > bloating the code. Compilers do a fairly good job of generating the specialisations appropriately, which leads to the observation that the inline keyword is kind of at the wrong level of granularity, as individual call sites are inlined, not functions. I built this program with a recent build of clang from SVN (left over from my work on Clang a few weeks back, as it happens) on a more powerful machine, and the difference narrowed: C quick-sort time elapsed: 1.89 Inline C quick-sort time elapsed: 1.54 This still seems significant though. > In any case I don't see how you're going to inline random database > functions. Those are the cases where large amounts of data are being > sorted. It's possible we sort small sets of data for internal reasons > very frequently but I don't recall it turning up at the top of the > profiles posted in recent days. Well, this point was raised in relation to the OP's patch, which does have a couple of simple comparators that look like good candidates for inlining. There are other really simple comparators around like that - one that I'm aware of off-hand is the one used to sort pg_stat_statements entries to find victims. It doesn't have to have major benefits to be worth undertaking - it only has to be worth the effort of its original implementation and ongoing maintenance. > Now a JIT that actually inlined random database functions into qsort > and optimized the result would be pretty cool. But it doesn't seem > like it's going to happen tomorrow... Yeah, that would be extremely cool. I'd guess that the main reason it isn't going to happen tomorrow is not so much technical, but because there would be about a dozen controversies involved in integrating an available, suitable JIT - how does it integrate with the build system, licensing, choice of implementation language, support on less popular platforms, how do package managers handle the dependency, can we trust the developers/steering committee of the JIT (Okay, so I'm really thinking about LLVM here), and so on. That's just off the top of my head. -- Peter Geoghegan http://www.2ndQuadrant.com/ PostgreSQL Development, 24x7 Support, Training and Services
I've processed the results of the tests with double sorting split which I've sheduled for buffering build. I've updated wiki page with them:
Raw results of query speed measues are in the attachment. There number of pages accesse is presented in dependency of table, buffering build and split method.
Note, that tests were run with not last version of fast build patch ("gist_fast_build-heikki-0.14.1.1.patch" was used). Therefore, build time with buffering can be better.
With best regards,
Alexander Korotkov.
Attachment
> ! /* > ! * Calculate delta between penalties of join "common entries" to > ! * different groups. > ! */ > ! for (i = 0; i < commonEntriesCount; i++) > { > ! double lower, > ! upper; > ! > ! box = DatumGetBoxP(entryvec->vector[commonEntries[i].index].key); > ! if (context.dim == 0) > ! { > ! lower = box->low.x; > ! upper = box->high.x; > ! } > ! else > ! { > ! lower = box->low.y; > ! upper = box->high.y; > ! } > ! commonEntries[i].delta = Abs(box_penalty(leftBox, box) - > ! box_penalty(rightBox, box)); > } 'lower' and 'upper' are not used for anything in the above. Is that just dead code that can be removed, or is there something missing that should be using them? -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
<div class="gmail_quote">On Thu, Sep 22, 2011 at 3:22 PM, Heikki Linnakangas <span dir="ltr"><<a href="mailto:heikki.linnakangas@enterprisedb.com"target="_blank">heikki.linnakangas@enterprisedb.com</a>></span> wrote:<br/><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><blockquoteclass="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">! /*<br /> ! * Calculate delta between penalties of join "common entries"to<br /> ! * different groups.<br /> ! */<br /> ! for (i = 0; i <commonEntriesCount; i++)<br /> {<br /> ! double lower,<br /> ! upper;<br /> !<br /> ! box = DatumGetBoxP(entryvec->vector[<u></u>commonEntries[i].index].key);<br/> ! if (context.dim == 0)<br/> ! {<br /> ! lower = box->low.x;<br /> ! upper = box->high.x;<br /> ! }<br /> ! else<br /> ! {<br /> ! lower = box->low.y;<br /> ! upper =box->high.y;<br /> ! }<br /> ! commonEntries[i].delta = Abs(box_penalty(leftBox,box) -<br /> ! box_penalty(rightBox,box));<br /> }<br /></blockquote><br /> 'lower' and 'upper' are not used for anythingin the above. Is that just dead code that can be removed, or is there something missing that should be using them?</blockquote>Yes,it's just dead code.</div><div class="gmail_quote"><br />------<br />With best regards,<br />AlexanderKorotkov. </div>
On Thu, Sep 22, 2011 at 3:31 PM, Alexander Korotkov <aekorotkov@gmail.com> wrote:
Patch without that dead code is attached.Yes, it's just dead code.On Thu, Sep 22, 2011 at 3:22 PM, Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote:'lower' and 'upper' are not used for anything in the above. Is that just dead code that can be removed, or is there something missing that should be using them?
------
With best regards,
Alexander Korotkov.
Attachment
On 22.09.2011 22:12, Alexander Korotkov wrote: > Patch without that dead code is attached. Thanks. Can you elaborate the consider-split algorithm? The criteria to select the new split over the previously selected one is this: > ! /* > ! * If ratio is acceptable, we should compare current split with > ! * previously selected one. If no split was selected then we select > ! * current anyway. Between splits of one dimension we search for > ! * minimal overlap (allowing negative values) and minimal ration > ! * (between same overlaps. We switch dimension if find less overlap > ! * (non-negative) or less range with same overlap. > ! */ > ! range = diminfo->upper - diminfo->lower; > ! overlap = ((leftUpper) - (rightLower)) / range; > ! if (context->first || > ! (context->dim == dimNum && > ! (overlap < context->overlap || > ! (overlap == context->overlap && ratio > context->ratio))) || > ! (context->dim != dimNum && > ! ((range > context->range && > ! non_negative(overlap) <= non_negative(context->overlap)) || > ! non_negative(overlap) < non_negative(context->overlap))) > ! ) > ! { Why are negative overlaps handled differently across dimensions and within the same dimension? Your considerSplit algorithm in the SYRCoSE 2011 paper doesn't seem to make that distinction. -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
On Tue, Oct 4, 2011 at 12:12 PM, Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote:
------
With best regards,
Alexander Korotkov.
Can you elaborate the consider-split algorithm? The criteria to select the new split over the previously selected one is this:! /*
! * If ratio is acceptable, we should compare current split with
! * previously selected one. If no split was selected then we select
! * current anyway. Between splits of one dimension we search for
! * minimal overlap (allowing negative values) and minimal ration
! * (between same overlaps. We switch dimension if find less overlap
! * (non-negative) or less range with same overlap.
! */
! range = diminfo->upper - diminfo->lower;
! overlap = ((leftUpper) - (rightLower)) / range;
! if (context->first ||
! (context->dim == dimNum &&
! (overlap < context->overlap ||
! (overlap == context->overlap && ratio > context->ratio))) ||
! (context->dim != dimNum &&
! ((range > context->range &&
! non_negative(overlap) <= non_negative(context->overlap)) ||
! non_negative(overlap) < non_negative(context->overlap)))
! )
! {
Why are negative overlaps handled differently across dimensions and within the same dimension? Your considerSplit algorithm in the SYRCoSE 2011 paper doesn't seem to make that distinction.
Yes, I've changed this behaviour after experiments on real-life datasets. On the datasets where data don't overlap themselfs (points are always such datasets), non-overlapping splits are frequently possible. In this case splits tends to be along one dimension, because most distant non-overlapping splits (i.e. having lowest negative overlap) appears to be in the same dimension as previous split. Therefore MBRs appear to be very prolonged along another dimension, that's bad for search. In order to prevent this behavour I've made transition to another dimension by non-nagative part of overlap and range. Using range as the split criteria makes MBRs more quadratic, and using non-negative part of overlap as priority criteria give to range chance to matter.
With best regards,
Alexander Korotkov.
On 04.10.2011 11:51, Alexander Korotkov wrote: > On Tue, Oct 4, 2011 at 12:12 PM, Heikki Linnakangas< > heikki.linnakangas@enterprisedb.com> wrote: > >> Can you elaborate the consider-split algorithm? The criteria to select the >> new split over the previously selected one is this: >> >>> ! /* >>> ! * If ratio is acceptable, we should compare current split >>> with >>> ! * previously selected one. If no split was selected then >>> we select >>> ! * current anyway. Between splits of one dimension we >>> search for >>> ! * minimal overlap (allowing negative values) and minimal >>> ration >>> ! * (between same overlaps. We switch dimension if find >>> less overlap >>> ! * (non-negative) or less range with same overlap. >>> ! */ >>> ! range = diminfo->upper - diminfo->lower; >>> ! overlap = ((leftUpper) - (rightLower)) / range; >>> ! if (context->first || >>> ! (context->dim == dimNum&& >>> ! (overlap< context->overlap || >>> ! (overlap == context->overlap&& ratio> >>> context->ratio))) || >>> ! (context->dim != dimNum&& >>> ! ((range> context->range&& >>> ! non_negative(overlap)<= >>> non_negative(context->overlap)**) || >>> ! non_negative(overlap)< >>> non_negative(context->overlap)**)) >>> ! ) >>> ! { >>> >> >> Why are negative overlaps handled differently across dimensions and within >> the same dimension? Your considerSplit algorithm in the SYRCoSE 2011 paper >> doesn't seem to make that distinction. > > Yes, I've changed this behaviour after experiments on real-life datasets. On > the datasets where data don't overlap themselfs (points are always such > datasets), non-overlapping splits are frequently possible. In this case > splits tends to be along one dimension, because most distant non-overlapping > splits (i.e. having lowest negative overlap) appears to be in the same > dimension as previous split. Therefore MBRs appear to be very prolonged > along another dimension, that's bad for search. In order to prevent this > behavour I've made transition to another dimension by non-nagative part of > overlap and range. Using range as the split criteria makes MBRs more > quadratic, and using non-negative part of overlap as priority criteria give > to range chance to matter. Ok. Could you phrase that as a code comment? Here's a version of the patch I've been working on. There's no functional changes, just a lot of moving things around, comment changes, etc. to hopefully make it more readable. -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
Attachment
On Tue, Oct 4, 2011 at 1:46 PM, Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote:
Thanks for your work on this patch. Patch with comment is attached.
Ok. Could you phrase that as a code comment?
Here's a version of the patch I've been working on. There's no functional changes, just a lot of moving things around, comment changes, etc. to hopefully make it more readable.
Thanks for your work on this patch. Patch with comment is attached.
------
With best regards,
Alexander Korotkov.
Attachment
On 04.10.2011 15:10, Alexander Korotkov wrote: > On Tue, Oct 4, 2011 at 1:46 PM, Heikki Linnakangas< > heikki.linnakangas@enterprisedb.com> wrote: > >> Ok. Could you phrase that as a code comment? >> >> Here's a version of the patch I've been working on. There's no functional >> changes, just a lot of moving things around, comment changes, etc. to >> hopefully make it more readable. > > Thanks for your work on this patch. Patch with comment is attached. Thanks, I incorporated that, and did a lot of other comment changes. I included the example you gave earlier on how the first phase of the algorithm works, in a comment. Please review, and if you have some test cases at hand, run them. I think this is ready for commit now. One more thing: > /* Allocate vectors for results */ > nbytes = (maxoff + 2) * sizeof(OffsetNumber); > v->spl_left = (OffsetNumber *) palloc(nbytes); > v->spl_right = (OffsetNumber *) palloc(nbytes); Why "maxoff + 2" ? Allocating a few extra bytes is obviously harmless, but I wonder if it was just a leftover from something. -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
Attachment
On Wed, Oct 5, 2011 at 11:37 AM, Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote:
------
With best regards,
Alexander Korotkov.
On 04.10.2011 15:10, Alexander Korotkov wrote:Thanks, I incorporated that, and did a lot of other comment changes. I included the example you gave earlier on how the first phase of the algorithm works, in a comment. Please review, and if you have some test cases at hand, run them. I think this is ready for commit now.On Tue, Oct 4, 2011 at 1:46 PM, Heikki Linnakangas<
heikki.linnakangas@enterprisedb.com> wrote:Ok. Could you phrase that as a code comment?
Here's a version of the patch I've been working on. There's no functional
changes, just a lot of moving things around, comment changes, etc. to
hopefully make it more readable.
Thanks for your work on this patch. Patch with comment is attached.
Comments looks good, thanks. I'm going to try also some datasets from rtreeportal.org
One more thing:/* Allocate vectors for results */
nbytes = (maxoff + 2) * sizeof(OffsetNumber);
v->spl_left = (OffsetNumber *) palloc(nbytes);
v->spl_right = (OffsetNumber *) palloc(nbytes);
Why "maxoff + 2" ? Allocating a few extra bytes is obviously harmless, but I wonder if it was just a leftover from something.
It was nested from old code. This extra bytes are useless in modern versions of PostgreSQL as we found while seg picksplit patch discussion. Modern version of seg picksplit doesn't contain them:
------
With best regards,
Alexander Korotkov.
Path without allocating extra bytes is attached.
I run some more detailed tests on geonames and two smaller datasets from rtreeportal.org.
Test tables are following:
1) test1 - geonames
2) test2 - California Roads, containing the MBRs of 2,249,727 streets
3) test3 - Tiger Streams, containing the MBRs of 194,971 streams
Scripts for test queries generation and execution are attached (scripts.tar.gz).
Build times are given in the following table:
New linear Double sorting
test1 277.889630 273.766355
test2 73.262561 75.114079
test3 4.251563 4.425139
As we can see, build times differ insignificantly.
Index sizes are given in the following table:
New linear Double sorting
test1 572383232 578977792
test2 179929088 178569216
test3 15409152 15532032
As we can see, index sizes differ insignificantly.
Data about index quality testing are in the table test_results of the following structure:
tablename - name of the queried table
count - actual count of rows matching query
nominal_count - count of rows matching query which test generator tried to provide (test generator not always can create test query which return exactly same amount of rows as it was expected)
strategy - split strategy: 1 - new liner split(current), 2 - double sorting(this patch)
hits - number of pages hits for query execution.
Raw data are in the attachment (test_result.sql.gz). Report is below.
test=# select tablename, nominal_count, round(avg(count),2) as avg_count, strategy, round(avg(hits),2) as avg_hits from test_results group by tablename, nominal_count, strategy order by tablename, nominal_count, strategy;
tablename | nominal_count | avg_count | strategy | avg_hits
-----------+---------------+-----------+----------+----------
test1 | 1 | 4.87 | 1 | 19.94
test1 | 1 | 4.87 | 2 | 6.46
test1 | 10 | 11.07 | 1 | 23.95
test1 | 10 | 11.07 | 2 | 7.36
test1 | 100 | 101.36 | 1 | 43.30
test1 | 100 | 101.36 | 2 | 10.19
test1 | 1000 | 998.70 | 1 | 86.48
test1 | 1000 | 998.70 | 2 | 24.21
test2 | 1 | 1.32 | 1 | 8.67
test2 | 1 | 1.32 | 2 | 5.99
test2 | 10 | 11.32 | 1 | 9.40
test2 | 10 | 11.32 | 2 | 6.71
test2 | 100 | 102.93 | 1 | 13.10
test2 | 100 | 102.93 | 2 | 9.02
test2 | 1000 | 999.67 | 1 | 32.16
test2 | 1000 | 999.67 | 2 | 23.51
test3 | 1 | 0.99 | 1 | 6.03
test3 | 1 | 0.99 | 2 | 4.32
test3 | 10 | 9.95 | 1 | 7.52
test3 | 10 | 9.95 | 2 | 5.09
test3 | 100 | 99.92 | 1 | 10.73
test3 | 100 | 99.92 | 2 | 7.73
test3 | 1000 | 999.75 | 1 | 27.47
test3 | 1000 | 999.75 | 2 | 22.44
(24 rows)
------
With best regards,
Alexander Korotkov.
Attachment
On 05.10.2011 15:59, Alexander Korotkov wrote: > Path without allocating extra bytes is attached. > I run some more detailed tests on geonames and two smaller datasets from > rtreeportal.org. Ok, thanks. Looks good to me now, so committed. -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
On Thu, Oct 6, 2011 at 11:06 AM, Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote:
------
With best regards,
Alexander Korotkov.
On 05.10.2011 15:59, Alexander Korotkov wrote:Ok, thanks. Looks good to me now, so committed.Path without allocating extra bytes is attached.
I run some more detailed tests on geonames and two smaller datasets from
rtreeportal.org.
Thanks. I'm going to continue work on application of this split method in following areas:
1) range types
2) seg contrib module
3) cube contrib module (there situation is not so easy, probably some heuristic of split method selection would be required)
Do you think that separation of some common parts of algorithm implemetation is resonable in order to avoid code duplication? For example, different application of algorithm could share function of splits enumeration along axis which takes pointer to consider_split as an argument.
------
With best regards,
Alexander Korotkov.
On 06.10.2011 11:22, Alexander Korotkov wrote: > Thanks. I'm going to continue work on application of this split method in > following areas: > 1) range types > 2) seg contrib module > 3) cube contrib module (there situation is not so easy, probably some > heuristic of split method selection would be required) > Do you think that separation of some common parts of algorithm implemetation > is resonable in order to avoid code duplication? For example, different > application of algorithm could share function of splits enumeration along > axis which takes pointer to consider_split as an argument. Dunno. I'd suggest that you just copy-paste the code for now, and we'll see how much duplication there is in the end. -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com