Re: Adding skip scan (including MDAM style range skip scan) to nbtree - Mailing list pgsql-hackers
From | Dmitry Dolgov |
---|---|
Subject | Re: Adding skip scan (including MDAM style range skip scan) to nbtree |
Date | |
Msg-id | 6loebaft2tpdwlvdzojuev5i5mcer7rituumiueocrsrripixa@pbfj47yl3rll Whole thread Raw |
In response to | Re: Adding skip scan (including MDAM style range skip scan) to nbtree (Peter Geoghegan <pg@bowt.ie>) |
Responses |
Re: Adding skip scan (including MDAM style range skip scan) to nbtree
|
List | pgsql-hackers |
> On Wed, Jun 26, 2024 at 03:16:07PM GMT, Peter Geoghegan wrote: > > Loose index scan is a far more specialized technique than skip scan. > It only applies within special scans that feed into a DISTINCT group > aggregate. Whereas my skip scan patch isn't like that at all -- it's > much more general. With my patch, nbtree has exactly the same contract > with the executor/core code as before. There are no new index paths > generated by the optimizer to make skip scan work, even. Skip scan > isn't particularly aimed at improving group aggregates (though the > benchmark I'll show happens to involve a group aggregate, simply > because the technique works best with large and expensive index > scans). I see that the patch is not supposed to deal with aggregates in any special way. But from what I understand after a quick review, skip scan is not getting applied to them if there are no quals in the query (in that case _bt_preprocess_keys returns before calling _bt_preprocess_array_keys). Yet such queries could benefit from skipping, I assume they still could be handled by the machinery introduced in this patch? > > Currently, there is an assumption that "there will be 10 primitive index scans > > per skipped attribute". Is any chance to use pg_stats.n_distinct? > > It probably makes sense to use pg_stats.n_distinct here. But how? > > If the problem is that we're too pessimistic, then I think that this > will usually (though not always) make us more pessimistic. Isn't that > the wrong direction to go in? (We're probably also too optimistic in > some cases, but being too pessimistic is a bigger problem in > practice.) > > For example, your test case involved 11 distinct values in each > column. The current approach of hard-coding 10 (which is just a > temporary hack) should actually make the scan look a bit cheaper than > it would if we used the true ndistinct. > > Another underlying problem is that the existing SAOP costing really > isn't very accurate, without skip scan -- that's a big source of the > pessimism with arrays/skipping. Why should we be able to get the true > number of primitive index scans just by multiplying together each > omitted prefix column's ndistinct? That approach is good for getting > the worst case, which is probably relevant -- but it's probably not a > very good assumption for the average case. (Though at least we can cap > the total number of primitive index scans to 1/3 of the total number > of pages in the index in btcostestimate, since we have guarantees > about the worst case as of Postgres 17.) Do I understand correctly, that the only way how multiplying ndistincts could produce too pessimistic results is when there is a correlation between distinct values? Can one benefit from the extended statistics here? And while we're at it, I think it would be great if the implementation will allow some level of visibility about the skip scan. From what I see, currently it's by design impossible for users to tell whether something was skipped or not. But when it comes to planning and estimates, maybe it's not a bad idea to let explain analyze show something like "expected number of primitive scans / actual number of primitive scans".
pgsql-hackers by date: