Re: Custom explain options - Mailing list pgsql-hackers
From | Konstantin Knizhnik |
---|---|
Subject | Re: Custom explain options |
Date | |
Msg-id | 57bc73d0-5f70-40bb-8d9e-ba4bc1f59d47@garret.ru Whole thread Raw |
In response to | Re: Custom explain options (Tomas Vondra <tomas.vondra@enterprisedb.com>) |
Responses |
Re: Custom explain options
|
List | pgsql-hackers |
The thing that was not clear to me is who decides what to prefetch, which code issues the prefetch requests etc. In the github links you shared I see it happens in the index AM code (in nbtsearch.c).
It is up to the particular plan node (seqscan, indexscan,...) which pages to prefetch.
That's interesting, because that's what my first prefetching patch did too - not the same way, ofc, but in the same layer. Simply because it seemed like the simplest way to do that. But the feedback was that's the wrong layer, and that it should happen in the executor. And I agree with that - the reasons are somewhere in the other thread.
I read the arguments in
Separating prefetch info in index scan descriptor is really good idea. It will be amazing to have generic prefetch mechanism for all indexes.
But unfortunately I do not understand how it is possible. The logic of index traversal is implemented inside AM. Executor doesn't know it.
For example for B-Tree scan we can prefetch:
- intermediate pages
- leave pages
- referenced by TID heap pages
Before we load next intermediate page, we do not know next leave pages.
And before we load next leave page, we can not find out TIDs from this page.
Another challenge - is how far we should prefetch (as far as I understand both your and our approach using dynamically extended prefetch window)
We definitely need prefetch for heap scan (it gives the most advantages in performance), for vacuum and also for pg_prewarm. Also I tried to implement it for custom indexes such as pg_vector. I still not sure whether it is possible to create some generic solution which will work for all indexes.Based on what I saw in the neon code, I think it should be possible for neon to use "my" approach too, but that only works for the index scans, ofc. Not sure what to do about the other places.
I have also tried to implement alternative approach for prefetch based on access statistic.
It comes from use case of seqscan of table with larger toasted records. So for each record we have to extract its TOAST data.
It is done using standard index scan, but unfortunately index prefetch doesn't help much here: there is usually just one TOAST segment and so prefetch just have no chance to do something useful. But as far as heap records are accessed sequentially, there is good chance that toast table will also be accessed mostly sequentially. So we just can count number of sequential requests to each relation and if ratio or seq/rand accesses is above some threshold we can prefetch next pages of this relation. This is really universal approach but ... working mostly for TOAST table.
As I already wrote - prefetch is done locally for each backend. And each backend has its own connection with page server. It can be changed in future when we implement multiplexing of page server connections. But right now prefetch is local. And certainly prefetch can improve performance only if we correctly predict subsequent page requests. If not - then page server does useless jobs and backend has to waity and consume all issues prefetch requests. This is why in prefetch implementation for most of nodes we start with minimal prefetch distance and then increase it. It allows to perform prefetch only for such queries where it is really efficient (OLAP) and doesn't degrade performance of simple OLTP queries.Not sure I understand what's so important about prefetches being "local" for each backend. I mean even in postgres each backend prefetches it's own buffers, no matter what the other backends do. Although, neon probably doesn't have the cross-backend sharing through shared buffers etc. right?
Sorry if my explanation was not clear:(
> I mean even in postgres each backend prefetches it's own buffers, no matter what the other backends do. This is exactly the difference. In Neon such approach doesn't work. Each backend maintains it's own prefetch ring. And if prefetched page was not actually received, then the whole pipe is lost. I.e. backend prefetched pages 1,5,10. Then it need to read page 2. So it has to consume responses for 1,5,10 and issue another request for page 2. Instead of improving speed we are just doing extra job. So each backend should prefetch only those pages which it is actually going to read. This is why prefetch approach used in Postgres for example for parallel bitmap heap scan doesn't work for Neon. If you do `posic_fadvise` then prefetched page is placed in OS cache and can be used by any parallel worker. But in Neon each parallel worker should be given its own range of pages to scan and prefetch only this pages.
I think that it is also problem without prefetch. There are cases where seqscan or bitmap heap scan are really much faster then IOS because last one has to perform a lot of visibility checks. Yes, certainly optimizer takes in account percent of all-visible pages. But with it is not tricial to adjust optimizer parameters so that it can really choose fastest plan.Well, my assumption was the following: prefetch is most efficient forOLAP queries. Although HTAP (hybrid transactional/analytical processing) is popular trend now, classical model is that analytic queries are performed on "historical" data, which was already proceeded by vacuum and all-visible bits were set in VM. May be this assumption is wrong but it seems to me that if most heap pages are not marked as all-visible, then optimizer should prefetch bitmap scan to index-only scan.
I think this assumption is generally reasonable, but it hinges on the assumption that OLAP queries have most indexes recently vacuumed and all-visible. I'm not sure it's wise to rely on that. Without prefetching it's not that important - the worst thing that would happen is that the IOS degrades into regular index-scan.
But withprefetching these plans can "invert" with respect to cost.
I'm not saying it's terrible or that IOS must have prefetching, but I
think it's something users may run into fairly often. And it led me to
rework the prefetching so that IOS can prefetch too ...
I think that inspecting VM for prefetch is really good idea.
Thanks! Very helpful. As I said, I ended up moving the prefetching to the executor. For indexscans I think it should be possible for neon to benefit from that (in a way, it doesn't need to do anything except for overriding what PrefetchBuffer does). Not sure about the other places where neon needs to prefetch, I don't have ambition to rework those.
Once your PR will be merged, I will rewrite Neon prefetch implementation fopr indexces using your approach.
pgsql-hackers by date: