Re: why do we need two snapshots per query? - Mailing list pgsql-hackers
From | Robert Haas |
---|---|
Subject | Re: why do we need two snapshots per query? |
Date | |
Msg-id | CA+TgmoYfg1DkGqnLAmmgtvhRoGXp11gX_WN_-VTkdwtmJ+0qXQ@mail.gmail.com Whole thread Raw |
In response to | Re: why do we need two snapshots per query? (Tom Lane <tgl@sss.pgh.pa.us>) |
Responses |
Re: why do we need two snapshots per query?
|
List | pgsql-hackers |
On Sun, Nov 13, 2011 at 12:57 PM, Tom Lane <tgl@sss.pgh.pa.us> wrote: >> Wait a minute. I can understand why you think it's a bad idea to >> preserve a snapshot across multiple protocol messages >> (parse/bind/execute), but why or how would it be a bad idea to keep >> the same snapshot between planning and execution when the whole thing >> is being done as a unit? You haven't offered any real justification >> for that position, > > It's not hard to come by: execution should proceed with the latest > available view of the database. The word "latest" doesn't seem very illuminating to me. If you take that to its (illogical) conclusion, that would mean that we ought to do everything under SnapshotNow - i.e. every time we fetch a tuple, use the "latest" available view of the database. It seems to me that you can wrap some logic around this - we shouldn't use a snapshot taken later than <event1> because <reason1>, and we shouldn't use one taken earlier than <event2> because <reason2>. It seems to me that the *latest* snapshot we could use would be one taken the instant before we did any calculation whose result might depend on our choice of snapshot. For example, if the query involves calculating pi out to 5000 decimal places (without looking at any tables) and then scanning for the matching value in some table column, we could do the whole calculation prior to taking a snapshot and then take the snapshot only when we start groveling through the table. That view would be "later" than the one we use now, and but still correct. On the other hand, it seems to me that the *earliest* snapshot we can use is one taken the instant after we receive the protocol message that tells us to execute the query. If we take it any sooner than that, we might fail to see as committed some transaction which was acknowledged before the user sent the message. Between those two extremes, it seems to me that when exactly the snapshot gets taken is an implementation detail. >> and it seems to me that if anything the semantics >> of such a thing are far *less* intuitive than it would be to do the >> whole thing under a single snapshot. > > In that case you must be of the opinion that extended query protocol > is a bad idea and we should get rid of it, and the same for prepared > plans of all types. What you're basically proposing is that simple > query mode will act differently from other ways of submitting a query, > and I don't think that's a good idea. I don't see why anything I said would indicate that we shouldn't have prepared plans. It is useful for users to have the option to parse and plan before execution - especially if they want to execute the same query repeatedly - and if they choose to make use of that functionality, then we and they will have to deal with the fact that things can change between plan time and execution time. If that means we miss some optimization opportunities, so be it. But we needn't deliver the semantics associated with the extended query protocol when the user isn't using it; and the next time we bump the protocol version we probably should give some thought to making sure that you only need to use the extended query protocol when you explicitly want to separate parse/plan from execution, and not just to get at some other functionality that we've only chosen to provided using the extended protocol. > It might be sane if planning > could be assumed to take zero time, but that's hardly true. I still maintain that the length of planning is irrelevant; more, if the planning and execution are happening in response to a single protocol message, then the semantics of the query need not (and perhaps even should not) depend on how much of that time is spent planning and how much is spent executing. >> I also think you are dismissing Simon's stable-expression-folding >> proposal far too lightly. I am not sure that the behavior he wants is >> safe given the current details of our implementation - or even with my >> patch; I suspect a little more than that is needed - but I am pretty >> certain it's the behavior that users want and expect, and we should be >> moving toward it, not away from it. I have seen a significant number >> of cases over the years where the query optimizer generated a bad plan >> because it did less constant-folding than the user expected. > > This is just FUD, unless you can point to specific examples where > Marti's patch won't fix it. If that patch crashes and burns for > some reason, then we should revisit this idea; but if it succeeds > it will cover more cases than plan-time constant folding could. I haven't reviewed the two patches in enough detail to have a clear understanding of which use cases each one does and does not cover. But, for example, you wrote this: tgl> As far as partitioning goes, the correct solution there tgl> is to move the partition selection to run-time, so we should not be tgl> contorting query semantics to make incremental performance improvements tgl> with the existing partitioning infrastructure. ...and I don't think I buy it. Certainly, being able to exclude partitions at runtime would be *extremely* valuable, but it's not a completely replacement for evaluating stable functions away prior to planning, because the latter approach allows the planner to see the function result and estimate the selectivity of that value specifically, which may lead to a much more accurate estimate and a completely different and far better plan. Logical purity is not, for me, a sufficient reason to throw that type of optimization out the window. Now that having been said, I'm pretty interested by what Marti is doing, too. > One of the reasons I don't want to go this direction is that it would > re-introduce causes of extended query protocol having poor performance > relative to simple protocol. That's not something that users find > intuitive or desirable, either. Insisting that we refuse to optimize the simple query protocol is the wrong solution to that problem. For what it's worth, the best result I was able to get with the patches I posted was about a 4% improvement on pgbench throughput (with 24-32 concurrent clients on a 32-core machine). So we're not talking about massively handicapping the extended query protocol. At the same time, we could sweat a lot more blood in other areas of the system for a lot less benefit. -- Robert Haas EnterpriseDB: http://www.enterprisedb.com The Enterprise PostgreSQL Company
pgsql-hackers by date: