Thread: Switching timeline over streaming replication
I've been working on the often-requested feature to handle timeline changes over streaming replication. At the moment, if you kill the master and promote a standby server, and you have another standby server that you'd like to keep following the new master server, you need a WAL archive in addition to streaming replication to make it cross the timeline change. Streaming replication will just error out. Having a WAL archive is usually a good idea in complex replication scenarios anyway, but it would be good to not require it. Attached is a WIP patch for that. It needs cleanup, but works. Protocol changes ---------------- When we invented the COPY-both mode, we left out any description of how to get out of that mode, simply stating that both ends "may then send CopyData messages until the connection is terminated". The patch makes it possible to revert back to regular processing, by sending a CopyDone message, like in normal Copy-in or Copy-out mode. Either end can take the initiative and send CopyDone, and after doing that may not send any more CopyDone messages. When both ends have sent a CopyDone message, and received a CopyDone message from the other end, the connection is out of Copy-mode, and the server finishes the command with a CommandComplete message. Another way to think of it is that when the server sends a CopyDone message, the connection switches from copy-both to Copy-in mode. And if the client sends a CopyDone message first, the connection goes from Copy-both to Copy-out mode, until the server ends the streaming from its end. New replication command: TIMELINE_HISTORY ----------------------------------------- To switch recovery target timeline, a standby needs the timeline history file (e.g 00000002.history) of the new timeline. The patch adds a new command to the set of commands accepted by walsender, to transmit a given timeline history file from master to slave. Walsender changes to stream a particular timeline ------------------------------------------------- The walsender now keeps track of exactly which timeline it is currently streaming; it's not necessarily the latest one anymore. The START_REPLICATION command is extended with a TIMELINE option that the client can use to request streaming from a particular timeline. If the client asks for a timeline that's not the current, but is part of the history of the server, the walsender knows to read from the correct WAL file that contains that. Also, the walsender knows where the server's history branched off from that timeline, and will only stream WAL up to that point. When that point is reached, it ends the streaming (with a CopyDone message), and prepares to accept a new replication command. Typically, the walreceiver will then ask to start streaming from the next timeline. Walreceiver changes ------------------- Previously, when the timeline reported by the server didn't match the current timeline in the standby, walreceiver simply errored out. Now, it requests for any missing timeline history files using the new TIMELINE_HISTORY command, and then tries to start replication from the current standby's timeline, even if that's older than the master's. When the end of the old timeline is reached, walreceiver sets a state variable in shared memory to indicate that, pings the the startup process, and waits for the startup process for new orders. The startup process can set receiveStart and timeline in shared memory and ping the walreceiver again, to get the walreceiver to restart streaming from the new starting point [1]. Before the startup process does that, it will scan pg_xlog for new timeline history files if recovery_target_timeline='latest'. It will find any new histrory files the walreceiver stored there, and switch over to the latest timeline just as it does with a WAL archive. Some parts of this patch are just refactoring that probably make sense regardless of the new functionality. For example, I split off the timeline history file related functions to a new file, timeline.c. That's not very much code, but it's fairly isolated, and xlog.c is massive, so I feel that anything that we can move off from xlog.c is a good thing. I also moved off the two functions RestoreArchivedFile() and ExecuteRecoveryCommand(), to a separate file. Those are also not much code, but are fairly isolated. If no-one objects to those changes, and the general direction this work is going to, I'm going split off those refactorings to separate patches and commit them separately. I also made the timeline history file a bit more detailed: instead of recording just the WAL segment where the timeline was changed, it now records the exact XLogRecPtr. That was required for the walsender to know the switchpoint, without having to parse the XLOG records (it reads and parses the history file, instead) [1] Initially, I tried to do this by simply letting walreceiver die and have the startup process launch a new walreceiver process that would reconnect, but it turned out to be hard to rapidly disconnect and connect, because the postmaster, which forks the walreceiver process, does not always have the same idea of whether the walreceiver is active as the startup process does. It would eventually be ok, thanks to timeouts, but would require polling. But not having to disconnect seems nicer, anyway - Heikki
Attachment
On Tuesday, September 11, 2012 10:53 PM Heikki Linnakangas wrote: > I've been working on the often-requested feature to handle timeline > changes over streaming replication. At the moment, if you kill the > master and promote a standby server, and you have another standby > server that you'd like to keep following the new master server, you > need a WAL archive in addition to streaming replication to make it > cross the timeline change. Streaming replication will just error out. > Having a WAL archive is usually a good idea in complex replication > scenarios anyway, but it would be good to not require it. Confirm my understanding of this feature: This feature is for case when standby-1 who is going to be promoted to master has archive mode 'on'. As in that case only its timeline will change. If above is right, then there can be other similar scenario's where it can be used: Scenario-1 (1 Master, 1 Stand-by) 1. Master (archive_mode=on) goes down. 2. Master again comes up 3. Stand-by tries to follow it Now in above scenario also due to timeline mismatch it gives error, but your patch should fix it. > > > Some parts of this patch are just refactoring that probably make sense > regardless of the new functionality. For example, I split off the > timeline history file related functions to a new file, timeline.c. > That's not very much code, but it's fairly isolated, and xlog.c is > massive, so I feel that anything that we can move off from xlog.c is a > good thing. I also moved off the two functions RestoreArchivedFile() > and ExecuteRecoveryCommand(), to a separate file. Those are also not > much code, but are fairly isolated. If no-one objects to those changes, > and the general direction this work is going to, I'm going split off > those refactorings to separate patches and commit them separately. > > I also made the timeline history file a bit more detailed: instead of > recording just the WAL segment where the timeline was changed, it now > records the exact XLogRecPtr. That was required for the walsender to > know the switchpoint, without having to parse the XLOG records (it > reads and parses the history file, instead) IMO separating timeline history file related functions to a new file is good. However I am not sure about splitting for RestoreArchivedFile() and ExecuteRecoveryCommand() into separate file. How about splitting for all Archive related functions: static void XLogArchiveNotify(const char *xlog); static void XLogArchiveNotifySeg(XLogSegNo segno); static bool XLogArchiveCheckDone(const char *xlog); static bool XLogArchiveIsBusy(const char *xlog); static void XLogArchiveCleanup(const char *xlog); .. .. In any case, it will be better if you can split it into multiple patches: 1. Having new functionality of "Switching timeline over streaming replication" 2. Refactoring related changes. It can make my testing and review for new feature patch little easier. With Regards, Amit Kapila.
> On Monday, September 24, 2012 9:08 PM md@rpzdesign.com wrote: > What a disaster waiting to happen. Maybe the only replication should be > master-master replication > so there is no need to sequence timelines or anything, all servers are > ready masters, no backups or failovers. > If you really do not want a master serving, then it should only be > handled in the routing > of traffic to that server and not the replication logic itself. The > only thing that ever came about > from failovers was the failure to turn over. The above is opinion > only. This feature is for users who want to use master-standby configurations. What do you mean by : "then it should only be handled in the routing of traffic to that server and not the replication logic itself." Do you have any idea other than proposed implementation or do you see any problem in currently proposed solution? > > On 9/24/2012 7:33 AM, Amit Kapila wrote: > > On Tuesday, September 11, 2012 10:53 PM Heikki Linnakangas wrote: > >> I've been working on the often-requested feature to handle timeline > >> changes over streaming replication. At the moment, if you kill the > >> master and promote a standby server, and you have another standby > >> server that you'd like to keep following the new master server, you > >> need a WAL archive in addition to streaming replication to make it > >> cross the timeline change. Streaming replication will just error > out. > >> Having a WAL archive is usually a good idea in complex replication > >> scenarios anyway, but it would be good to not require it. > > Confirm my understanding of this feature: > > > > This feature is for case when standby-1 who is going to be promoted > to > > master has archive mode 'on'. > > As in that case only its timeline will change. > > > > If above is right, then there can be other similar scenario's where > it can > > be used: > > > > Scenario-1 (1 Master, 1 Stand-by) > > 1. Master (archive_mode=on) goes down. > > 2. Master again comes up > > 3. Stand-by tries to follow it > > > > Now in above scenario also due to timeline mismatch it gives error, > but your > > patch should fix it. > > > > > >> > >> Some parts of this patch are just refactoring that probably make > sense > >> regardless of the new functionality. For example, I split off the > >> timeline history file related functions to a new file, timeline.c. > >> That's not very much code, but it's fairly isolated, and xlog.c is > >> massive, so I feel that anything that we can move off from xlog.c is > a > >> good thing. I also moved off the two functions RestoreArchivedFile() > >> and ExecuteRecoveryCommand(), to a separate file. Those are also not > >> much code, but are fairly isolated. If no-one objects to those > changes, > >> and the general direction this work is going to, I'm going split off > >> those refactorings to separate patches and commit them separately. > >> > >> I also made the timeline history file a bit more detailed: instead > of > >> recording just the WAL segment where the timeline was changed, it > now > >> records the exact XLogRecPtr. That was required for the walsender to > >> know the switchpoint, without having to parse the XLOG records (it > >> reads and parses the history file, instead) > > IMO separating timeline history file related functions to a new file > is > > good. > > However I am not sure about splitting for RestoreArchivedFile() and > > ExecuteRecoveryCommand() into separate file. > > How about splitting for all Archive related functions: > > static void XLogArchiveNotify(const char *xlog); > > static void XLogArchiveNotifySeg(XLogSegNo segno); > > static bool XLogArchiveCheckDone(const char *xlog); > > static bool XLogArchiveIsBusy(const char *xlog); > > static void XLogArchiveCleanup(const char *xlog); > > .. > > .. > > > > In any case, it will be better if you can split it into multiple > patches: > > 1. Having new functionality of "Switching timeline over streaming > > replication" > > 2. Refactoring related changes. > > > > It can make my testing and review for new feature patch little > easier. > > > > With Regards, > > Amit Kapila. > > > > > >
On 24.09.2012 16:33, Amit Kapila wrote: > On Tuesday, September 11, 2012 10:53 PM Heikki Linnakangas wrote: >> I've been working on the often-requested feature to handle timeline >> changes over streaming replication. At the moment, if you kill the >> master and promote a standby server, and you have another standby >> server that you'd like to keep following the new master server, you >> need a WAL archive in addition to streaming replication to make it >> cross the timeline change. Streaming replication will just error out. >> Having a WAL archive is usually a good idea in complex replication >> scenarios anyway, but it would be good to not require it. > > Confirm my understanding of this feature: > > This feature is for case when standby-1 who is going to be promoted to > master has archive mode 'on'. No. This is for the case where there is no WAL archive. archive_mode='off' on all servers. Or to be precise, you can also have a WAL archive, but this patch doesn't affect that in any way. This is strictly about streaming replication. > As in that case only its timeline will change. The timeline changes whenever you promote a standby. It's not related to whether you have a WAL archive or not. > If above is right, then there can be other similar scenario's where it can > be used: > > Scenario-1 (1 Master, 1 Stand-by) > 1. Master (archive_mode=on) goes down. > 2. Master again comes up > 3. Stand-by tries to follow it > > Now in above scenario also due to timeline mismatch it gives error, but your > patch should fix it. If the master simply crashes or is shut down, and then restarted, the timeline doesn't change. The standby will reconnect / poll the archive, and sync up just fine, even without this patch. > However I am not sure about splitting for RestoreArchivedFile() and > ExecuteRecoveryCommand() into separate file. > How about splitting for all Archive related functions: > static void XLogArchiveNotify(const char *xlog); > static void XLogArchiveNotifySeg(XLogSegNo segno); > static bool XLogArchiveCheckDone(const char *xlog); > static bool XLogArchiveIsBusy(const char *xlog); > static void XLogArchiveCleanup(const char *xlog); Hmm, sounds reasonable. > In any case, it will be better if you can split it into multiple patches: > 1. Having new functionality of "Switching timeline over streaming > replication" > 2. Refactoring related changes. > > It can make my testing and review for new feature patch little easier. Yep, I'll go ahead and split the patch. Thanks! - Heikki
On Tuesday, September 25, 2012 12:39 PM Heikki Linnakangas wrote: > On 24.09.2012 16:33, Amit Kapila wrote: > > On Tuesday, September 11, 2012 10:53 PM Heikki Linnakangas wrote: > >> I've been working on the often-requested feature to handle timeline > >> changes over streaming replication. At the moment, if you kill the > >> master and promote a standby server, and you have another standby > >> server that you'd like to keep following the new master server, you > >> need a WAL archive in addition to streaming replication to make it > >> cross the timeline change. Streaming replication will just error > out. > >> Having a WAL archive is usually a good idea in complex replication > >> scenarios anyway, but it would be good to not require it. > > > > Confirm my understanding of this feature: > > > > This feature is for case when standby-1 who is going to be promoted > to > > master has archive mode 'on'. > > No. This is for the case where there is no WAL archive. > archive_mode='off' on all servers. > > Or to be precise, you can also have a WAL archive, but this patch > doesn't affect that in any way. This is strictly about streaming > replication. > > > As in that case only its timeline will change. > > The timeline changes whenever you promote a standby. It's not related > to > whether you have a WAL archive or not. Yes that is correct. I thought timeline change happens only when somebody does PITR. Can you please tell me why we change timeline after promotion, because the original Timeline concept was for PITR and I am not able to trace from code the reason why on promotion it is required? > > > If above is right, then there can be other similar scenario's where > it can > > be used: > > > > Scenario-1 (1 Master, 1 Stand-by) > > 1. Master (archive_mode=on) goes down. > > 2. Master again comes up > > 3. Stand-by tries to follow it > > > > Now in above scenario also due to timeline mismatch it gives error, > but your > > patch should fix it. > > If the master simply crashes or is shut down, and then restarted, the > timeline doesn't change. The standby will reconnect / poll the archive, > and sync up just fine, even without this patch. How about when Master does PITR when it comes again? With Regards, Amit Kapila.
On 25.09.2012 14:10, Amit Kapila wrote: > On Tuesday, September 25, 2012 12:39 PM Heikki Linnakangas wrote: >> On 24.09.2012 16:33, Amit Kapila wrote: >>> On Tuesday, September 11, 2012 10:53 PM Heikki Linnakangas wrote: >>>> I've been working on the often-requested feature to handle timeline >>>> changes over streaming replication. At the moment, if you kill the >>>> master and promote a standby server, and you have another standby >>>> server that you'd like to keep following the new master server, you >>>> need a WAL archive in addition to streaming replication to make it >>>> cross the timeline change. Streaming replication will just error >> out. >>>> Having a WAL archive is usually a good idea in complex replication >>>> scenarios anyway, but it would be good to not require it. >>> >>> Confirm my understanding of this feature: >>> >>> This feature is for case when standby-1 who is going to be promoted >> to >>> master has archive mode 'on'. >> >> No. This is for the case where there is no WAL archive. >> archive_mode='off' on all servers. >> >> Or to be precise, you can also have a WAL archive, but this patch >> doesn't affect that in any way. This is strictly about streaming >> replication. >> >>> As in that case only its timeline will change. >> >> The timeline changes whenever you promote a standby. It's not related >> to >> whether you have a WAL archive or not. > > Yes that is correct. I thought timeline change happens only when somebody > does PITR. > Can you please tell me why we change timeline after promotion, because the > original > Timeline concept was for PITR and I am not able to trace from code the > reason > why on promotion it is required? Bumping the timeline helps to avoid confusion if, for example, the master crashes, and the standby isn't fully in sync with it. In that situation, there are some WAL records in the master that are not in the standby, so promoting the standby is effectively the same as doing PITR. If you promote the standby, and later try to turn the old master into a standby server that connects to the new master, things will go wrong. Assigning the new master a new timeline ID helps the system and the administrator to notice that. It's not bulletproof, for example you can easily avoid the timeline change if you just remove recovery.conf and restart the server, but the timelines help to manage such situations. >>> If above is right, then there can be other similar scenario's where >> it can >>> be used: >>> >>> Scenario-1 (1 Master, 1 Stand-by) >>> 1. Master (archive_mode=on) goes down. >>> 2. Master again comes up >>> 3. Stand-by tries to follow it >>> >>> Now in above scenario also due to timeline mismatch it gives error, >> but your >>> patch should fix it. >> >> If the master simply crashes or is shut down, and then restarted, the >> timeline doesn't change. The standby will reconnect / poll the archive, >> and sync up just fine, even without this patch. > > How about when Master does PITR when it comes again? Then the timeline will be bumped and this patch will be helpful. Assuming the standby is behind the point in time that the master was recovered to, it will be able to follow the master to the new timeline. - Heikki
On 25.09.2012 10:08, Heikki Linnakangas wrote: > On 24.09.2012 16:33, Amit Kapila wrote: >> In any case, it will be better if you can split it into multiple patches: >> 1. Having new functionality of "Switching timeline over streaming >> replication" >> 2. Refactoring related changes. >> >> It can make my testing and review for new feature patch little easier. > > Yep, I'll go ahead and split the patch. Thanks! Ok, here you go. xlog-c-split-1.patch contains the refactoring of existing code, with no user-visible changes. streaming-tli-switch-2.patch applies over xlog-c-split-1.patch, and contains the new functionality. - Heikki
Attachment
> On Tuesday, September 25, 2012 6:29 PM Heikki Linnakangas wrote: > On 25.09.2012 10:08, Heikki Linnakangas wrote: > > On 24.09.2012 16:33, Amit Kapila wrote: > >> In any case, it will be better if you can split it into multiple > patches: > >> 1. Having new functionality of "Switching timeline over streaming > >> replication" > >> 2. Refactoring related changes. > >> > >> It can make my testing and review for new feature patch little > easier. > > > > Yep, I'll go ahead and split the patch. Thanks! > > Ok, here you go. xlog-c-split-1.patch contains the refactoring of > existing code, with no user-visible changes. > streaming-tli-switch-2.patch applies over xlog-c-split-1.patch, and > contains the new functionality. Thanks, it will make my review easier than previous. With Regards, Amit Kapila.
Amit: At some point, every master - slave replicator gets to the point where they need to start thinking about master-master replication. Instead of getting stuck in the weeds to finally realize that master-master is the ONLY way to go, many developers do not start out planning for master - master, but they should, out of habit. You can save yourself a lot of grief just be starting with master-master architecture. But you don't have to USE it, you can just not send WRITE traffic to the servers that you do not want to WRITE to, but all of them should be WRITE servers. That way, the only timeline you ever need is your decision to send WRITE traffic request to them, but there is nothing that prevents you from running MASTER - MASTER all the time and skip the whole slave thing entirely. At this point, I think synchronous replication is only for immediate local replication needs and async for all the master - master stuff. cheers, marco On 9/24/2012 9:44 PM, Amit Kapila wrote: >> On Monday, September 24, 2012 9:08 PM md@rpzdesign.com wrote: >> What a disaster waiting to happen. Maybe the only replication should be >> master-master replication >> so there is no need to sequence timelines or anything, all servers are >> ready masters, no backups or failovers. >> If you really do not want a master serving, then it should only be >> handled in the routing >> of traffic to that server and not the replication logic itself. The >> only thing that ever came about >> from failovers was the failure to turn over. The above is opinion >> only. > This feature is for users who want to use master-standby configurations. > > What do you mean by : > "then it should only be handled in the routing of traffic to that server > and not the replication logic itself." > > Do you have any idea other than proposed implementation or do you see any > problem in currently proposed solution? > > >> On 9/24/2012 7:33 AM, Amit Kapila wrote: >>> On Tuesday, September 11, 2012 10:53 PM Heikki Linnakangas wrote: >>>> I've been working on the often-requested feature to handle timeline >>>> changes over streaming replication. At the moment, if you kill the >>>> master and promote a standby server, and you have another standby >>>> server that you'd like to keep following the new master server, you >>>> need a WAL archive in addition to streaming replication to make it >>>> cross the timeline change. Streaming replication will just error >> out. >>>> Having a WAL archive is usually a good idea in complex replication >>>> scenarios anyway, but it would be good to not require it. >>> Confirm my understanding of this feature: >>> >>> This feature is for case when standby-1 who is going to be promoted >> to >>> master has archive mode 'on'. >>> As in that case only its timeline will change. >>> >>> If above is right, then there can be other similar scenario's where >> it can >>> be used: >>> >>> Scenario-1 (1 Master, 1 Stand-by) >>> 1. Master (archive_mode=on) goes down. >>> 2. Master again comes up >>> 3. Stand-by tries to follow it >>> >>> Now in above scenario also due to timeline mismatch it gives error, >> but your >>> patch should fix it. >>> >>> >>>> Some parts of this patch are just refactoring that probably make >> sense >>>> regardless of the new functionality. For example, I split off the >>>> timeline history file related functions to a new file, timeline.c. >>>> That's not very much code, but it's fairly isolated, and xlog.c is >>>> massive, so I feel that anything that we can move off from xlog.c is >> a >>>> good thing. I also moved off the two functions RestoreArchivedFile() >>>> and ExecuteRecoveryCommand(), to a separate file. Those are also not >>>> much code, but are fairly isolated. If no-one objects to those >> changes, >>>> and the general direction this work is going to, I'm going split off >>>> those refactorings to separate patches and commit them separately. >>>> >>>> I also made the timeline history file a bit more detailed: instead >> of >>>> recording just the WAL segment where the timeline was changed, it >> now >>>> records the exact XLogRecPtr. That was required for the walsender to >>>> know the switchpoint, without having to parse the XLOG records (it >>>> reads and parses the history file, instead) >>> IMO separating timeline history file related functions to a new file >> is >>> good. >>> However I am not sure about splitting for RestoreArchivedFile() and >>> ExecuteRecoveryCommand() into separate file. >>> How about splitting for all Archive related functions: >>> static void XLogArchiveNotify(const char *xlog); >>> static void XLogArchiveNotifySeg(XLogSegNo segno); >>> static bool XLogArchiveCheckDone(const char *xlog); >>> static bool XLogArchiveIsBusy(const char *xlog); >>> static void XLogArchiveCleanup(const char *xlog); >>> .. >>> .. >>> >>> In any case, it will be better if you can split it into multiple >> patches: >>> 1. Having new functionality of "Switching timeline over streaming >>> replication" >>> 2. Refactoring related changes. >>> >>> It can make my testing and review for new feature patch little >> easier. >>> With Regards, >>> Amit Kapila. >>> >>> >>> >
On Tue, Sep 25, 2012 at 11:01 AM, md@rpzdesign.com <md@rpzdesign.com> wrote: > Amit: > > At some point, every master - slave replicator gets to the point where they > need > to start thinking about master-master replication. Even in a master-master system, the ability to cleanly swap leaders managing a member of the master-master cluster is very useful. This patch can make writing HA software for Postgres a lot less ridiculous. > Instead of getting stuck in the weeds to finally realize that master-master > is the ONLY way > to go, many developers do not start out planning for master - master, but > they should, out of habit. > > You can save yourself a lot of grief just be starting with master-master > architecture. I've seen more projects get stuck spinning their wheels on the one Master-Master system to rule them all then succeed and move on. It doesn't help that master-master does not have a single definition, and different properties are possible with different logical models, too, so that pervades its way up to the language layer. As-is, managing single-master HA Postgres is a huge pain without this patch. If there is work to be done on master-master, the logical replication and event trigger work are probably more relevant, and I know the authors of those projects are keen to make it more feasible to experiment. -- fdr
On 09/25/12 11:01 AM, md@rpzdesign.com wrote: > > > At some point, every master - slave replicator gets to the point where > they need > to start thinking about master-master replication. master-master and transactional integrity are mutually exclusive, except perhaps in special cases like Oracle RAC, where the masters share a coherent cache and implement global locks. -- john r pierce N 37, W 122 santa cruz ca mid-left coast
John: Who has the money for oracle RAC or funding arrogant bastard Oracle CEO Ellison to purchase another island? Postgres needs CHEAP, easy to setup, self healing, master-master-master-master and it needs it yesterday. I was able to patch the 9.2.0 code base in 1 day and change my entire architecture strategy for replication into self healing async master-master-master and the tiniest bit of sharding code imaginable That is why I suggest something to replace OIDs with ROIDs for replication ID. (CREATE TABLE with ROIDS) I implement ROIDs as a uniform design pattern for the table structures. Synchronous replication maybe between 2 local machines if absolutely no local hardware failure is acceptable, but cheap, scaleable synchronous, TRANSACTIONAL, master-master-master-master is a real tough slog. I could implement global locks in the external replication layer if I choose, but there are much easier ways in routing requests thru the load balancer and request sharding than trying to manage global locks across the WAN. Good luck with your HA patch for Postgres. Thanks for all of the responses! You guys are 15 times more active than the MySQL developer group, likely because they do not have a single db engine that meets all the requirements like PG. marco On 9/25/2012 5:10 PM, John R Pierce wrote: > On 09/25/12 11:01 AM, md@rpzdesign.com wrote: >> >> >> At some point, every master - slave replicator gets to the point >> where they need >> to start thinking about master-master replication. > > master-master and transactional integrity are mutually exclusive, > except perhaps in special cases like Oracle RAC, where the masters > share a coherent cache and implement global locks. > > > > >
> I was able to patch the 9.2.0 code base in 1 day and change my entire > architecture strategy for replication > into self healing async master-master-master and the tiniest bit of > sharding code imaginable Sounds cool. Do you have a fork available on Github? I'll try it out. -- Josh Berkus PostgreSQL Experts Inc. http://pgexperts.com
> Yes that is correct. I thought timeline change happens only when somebody > does PITR. > Can you please tell me why we change timeline after promotion, because the > original > Timeline concept was for PITR and I am not able to trace from code the > reason > why on promotion it is required? The idea behind the timeline switch is to prevent a server from subscribing to a master which is actually behind it. For example, consider this sequence: 1. M1->async->S1 2. M1 is at xid 2001 and fails. 3. S1 did not receive transaction 2001 and is at xid 2000 4. S1 is promoted. 5. S1 processed an new, different transaction 2001 6. M1 is repaired and brought back up 7. M1 is subscribed to S1 8. M1 is now corrupt. That's why we need the timeline switch. -- Josh Berkus PostgreSQL Experts Inc. http://pgexperts.com
Josh: The good part is you are the first person to ask for a copy and I will send you the hook code that I have and you can be a good sport and put it on GitHub, that is great, you can give us both credit for a joint effort, I do the code, you put it GitHub. The not so good part is that the community has a bunch of other trigger work and other stuff going on, so there was not much interest in non-WAL replication hook code. I do not have time to debate implementation nor wait for release of 9.3 with my needs not met, so I will just keep patching the hook code into whatever release code base comes along. The bad news is that I have not implemented the logic of the external replication daemon. The other good and bad news is that you are free to receive the messages from the hook code thru the unix socket and implement replication any way you want and the bad news is that you are free to IMPLEMENT replication any way you want. I am going to implement master-master-master-master SELF HEALING replication, but that is just my preference. Should take about a week to get it operational and another week to see how it works in my geographically dispersed servers in the cloud. Send me a note if it is ok to send you a zip file with the source code files that I touched in the 9.2 code base so you can shove it up on GitHub. Cheers, marco On 9/26/2012 6:48 PM, Josh Berkus wrote: >> I was able to patch the 9.2.0 code base in 1 day and change my entire >> architecture strategy for replication >> into self healing async master-master-master and the tiniest bit of >> sharding code imaginable > Sounds cool. Do you have a fork available on Github? I'll try it out. >
On Thursday, September 27, 2012 6:30 AM Josh Berkus wrote: > > Yes that is correct. I thought timeline change happens only when > somebody > > does PITR. > > Can you please tell me why we change timeline after promotion, > because the > > original > > Timeline concept was for PITR and I am not able to trace from code > the > > reason > > why on promotion it is required? > > The idea behind the timeline switch is to prevent a server from > subscribing to a master which is actually behind it. For example, > consider this sequence: > > 1. M1->async->S1 > 2. M1 is at xid 2001 and fails. > 3. S1 did not receive transaction 2001 and is at xid 2000 > 4. S1 is promoted. > 5. S1 processed an new, different transaction 2001 > 6. M1 is repaired and brought back up > 7. M1 is subscribed to S1 > 8. M1 is now corrupt. > > That's why we need the timeline switch. Thanks. I understood this point, but currently in documentation of Timelines, this usecase is not documented (Section 24.3.5). With Regards, Amit Kapila.
On 09/26/2012 01:02 AM, md@rpzdesign.com wrote: > John: > > Who has the money for oracle RAC or funding arrogant bastard Oracle > CEO Ellison to purchase another island? > > Postgres needs CHEAP, easy to setup, self healing, > master-master-master-master and it needs it yesterday. > > I was able to patch the 9.2.0 code base in 1 day and change my entire > architecture strategy for replication > into self healing async master-master-master and the tiniest bit of > sharding code imaginable Tell us about the compromises you had to make. It is an established fact that you can either have it replicate fast and loose or slow and correct. In the fast and loose case you have to be ready to do a lot of mopping-up in case of conflicts. > That is why I suggest something to replace OIDs with ROIDs for > replication ID. (CREATE TABLE with ROIDS) > I implement ROIDs as a uniform design pattern for the table structures. > > Synchronous replication maybe between 2 local machines if absolutely > no local > hardware failure is acceptable, but cheap, scaleable synchronous, Scaleable / synchronous is probably doable, if we are ready to take the initial performance hit of lock propagation. > TRANSACTIONAL, master-master-master-master is a real tough slog. > > I could implement global locks in the external replication layer if I > choose, but there are much easier ways in routing > requests thru the load balancer and request sharding than trying to > manage global locks across the WAN. > > Good luck with your HA patch for Postgres. > > Thanks for all of the responses! > > You guys are 15 times more active than the MySQL developer group, > likely because > they do not have a single db engine that meets all the requirements > like PG. > > marco > > On 9/25/2012 5:10 PM, John R Pierce wrote: >> On 09/25/12 11:01 AM, md@rpzdesign.com wrote: >>> >>> >>> At some point, every master - slave replicator gets to the point >>> where they need >>> to start thinking about master-master replication. >> >> master-master and transactional integrity are mutually exclusive, >> except perhaps in special cases like Oracle RAC, where the masters >> share a coherent cache and implement global locks. >> >> >> >> >> > > >
On 9/26/12 6:17 PM, md@rpzdesign.com wrote: > Josh: > > The good part is you are the first person to ask for a copy > and I will send you the hook code that I have and you can be a good sport > and put it on GitHub, that is great, you can give us both credit for a > joint effort, I do the code, > you put it GitHub. Well, I think it just makes sense for you to put it up somewhere public so that folks can review it; if not Github, then somewhere else. If it's useful and well-written, folks will be interested. -- Josh Berkus PostgreSQL Experts Inc. http://pgexperts.com
On 27-09-2012 01:30, Amit Kapila wrote: > I understood this point, but currently in documentation of Timelines, this usecase is not documented (Section 24.3.5). > Timeline documentation was written during PITR implementation. There wasn't SR yet. AFAICS it doesn't cite SR but is sufficiently generic (it use 'wal records' term to explain the feature). Feel free to reword those paragraphs mentioning SR. -- Euler Taveira de Oliveira - Timbira http://www.timbira.com.br/ PostgreSQL: Consultoria, Desenvolvimento, Suporte24x7 e Treinamento
> On Tuesday, September 25, 2012 6:29 PM Heikki Linnakangas wrote: > On 25.09.2012 10:08, Heikki Linnakangas wrote: > > On 24.09.2012 16:33, Amit Kapila wrote: > >> In any case, it will be better if you can split it into multiple > patches: > >> 1. Having new functionality of "Switching timeline over streaming > >> replication" > >> 2. Refactoring related changes. > Ok, here you go. xlog-c-split-1.patch contains the refactoring of existing code, with no user-visible changes. > streaming-tli-switch-2.patch applies over xlog-c-split-1.patch, and contains the new functionality. Please find the initial review of the patch. Still more review is pending, but I thought whatever is done I shall post Basic stuff: ---------------------- - Patch applies OK - Compiles cleanly with no warnings - Regression tests pass. - Documentation changes are mostly fine. - Basic replication tests works. Testing --------- Start primary server Start standby server Start cascade standby server Stopped the primary server Promoted the standby server with ./pg_ctl -D data_repl promote In postgresql.conf file archive_mode = off The following logs are observing in the cascade standby server. LOG: replication terminated by primary server DETAIL: End of WAL reached on timeline 1 LOG: walreceiver ended streaming and awaits new instructions LOG: record with zero length at 0/17E3888 LOG: re-handshaking at position 0/1000000 on tli 1 LOG: fetching timeline history file for timeline 2 from primary server LOG: replication terminated by primary server DETAIL: End of WAL reached on timeline 1 LOG: walreceiver ended streaming and awaits new instructions LOG: re-handshaking at position 0/1000000 on tli 1 LOG: replication terminated by primary server DETAIL: End of WAL reached on timeline 1 In postgresql.conf file archive_mode = on The following logs are observing in the cascade standby server. LOG: replication terminated by primary server DETAIL: End of WAL reached on timeline 1 LOG: walreceiver ended streaming and awaits new instructions sh: /home/amit/installation/bin/data_sub/pg_xlog/archive_status/0000000100000000 00000002: No such file or directory LOG: record with zero length at 0/20144B8 sh: /home/amit/installation/bin/data_sub/pg_xlog/archive_status/0000000100000000 00000002: No such file or directory LOG: re-handshaking at position 0/2000000 on tli 1 LOG: fetching timeline history file for timeline 2 from primary server LOG: replication terminated by primary server DETAIL: End of WAL reached on timeline 1 LOG: walreceiver ended streaming and awaits new instructions sh: /home/amit/installation/bin/data_sub/pg_xlog/archive_status/0000000100000000 00000002: No such file or directory sh: /home/amit/installation/bin/data_sub/pg_xlog/archive_status/0000000100000000 00000002: No such file or directory LOG: re-handshaking at position 0/2000000 on tli 1 LOG: replication terminated by primary server DETAIL: End of WAL reached on timeline 1 LOG: walreceiver ended streaming and awaits new instructions Verified that files are present in respective directories. Code Review ---------------- 1. In function readTimeLineHistory(), two mechanisms are used to fetch timeline from history file + sscanf(fline,"%u\t%X/%X", &tli, &switchpoint_hi, &switchpoint_lo); + + /* expect a numeric timeline ID as first field of line */ + tli = (TimeLineID) strtoul(ptr, &endptr, 0); If we use new mechanism, it will not be able to detect erroras it is doing in current case. 2. In function readTimeLineHistory(), + fd = AllocateFile(path, "r"); + if (fd == NULL) + { + if (errno != ENOENT) + ereport(FATAL, + (errcode_for_file_access(), + errmsg("could not open file \"%s\": %m", path))); + /* Not there, so assume no parents */ + return list_make1_int((int) targetTLI); + } still return list_make1_int((int) targetTLI); is used. 3. Function timelineOfPointInHistory(), should return the timeline of recptr passed to it. a. is it okay to decide based on xlog recordpointer that which timeline it belongs to, as different timelines can have same xlog recordpointer? b. it seems from logic that it will returntimeline previous to the timeline of recptr passed. For example if the timeline 3's switchpoint is equal to recptr passed then it will return timeline 2. 4. In writeTimeLineHistory function variable endTLI is never used. 5. In header of function writeTimeLineHistory(), can give explanation about XLogRecPtr switchpoint 6. @@ -6869,11 +5947,35 @@ StartupXLOG(void) */ if (InArchiveRecovery) { + char reason[200]; + + /* + * Write comment to history file to explain why and where timeline + * changed. Comment varies according to the recovery target used. + */ + if (recoveryTarget == RECOVERY_TARGET_XID) + snprintf(reason, sizeof(reason), + "%s transaction %u", + recoveryStopAfter ? "after" : "before", + recoveryStopXid); In the comment above this line you mentioned why and where timeline changed. However in the reason field only specifies about where part. 7. + * Returns the redo pointer of the "previous" checkpoint. +GetOldestRestartPoint(XLogRecPtr *oldrecptr, TimeLineID *oldtli) +{ + if (InRedo) + { + LWLockAcquire(ControlFileLock, LW_SHARED); + *oldrecptr = ControlFile->checkPointCopy.redo; + *oldtli = ControlFile->checkPointCopy.ThisTimeLineID; + LWLockRelease(ControlFileLock); + } a. In this function, is it required to take ControlFileLock as earlier also there was no lock to protect this read when it get called from RestoreArchivedFile, and I think at this point no one else can modify these values. However for code consistency purpose like whenever or wherever read the controlfile values, read it with read lock. b. As per your comment it should have returned "previous" checkpoint, however the function returns values of latest checkpoint. 8. In function writeTimeLineHistoryFile(), will it not be better to directly write rather than to later do pg_fsync(). as it is just one time write. 9. +XLogRecPtr +timeLineSwitchPoint(XLogRecPtr startpoint, TimeLineID tli) .. .. + * starting point. This is because the client can legimately spelling of legitimately needs to be corrected. 10.+XLogRecPtr +timeLineSwitchPoint(XLogRecPtr startpoint, TimeLineID tli) .. .. + if (tli < ThisTimeLineID) + { + if (!nexttle) + elog(ERROR, "could not find history entry for child of timeline %u", tli); /* shouldn't happen */ + } I don't understand the meaning of the above check, as I think this situation can occur when this function gets called from StartReplication, because always tli sent by standby to new master will be less than ThisTimeLineID and it can be first in list. Documentation --------------- 1. In explanation of TIMELINE_HISTORY: Filename of the timeline history file. This is always of the form [insert correct example here]. Give example. 2. In protocol.sgml change, I feel better explain when the COPYDONE message will be initiated. With Regards, Amit Kapila.
> On Friday, September 28, 2012 6:38 PM Amit Kapila wrote: > On Tuesday, September 25, 2012 6:29 PM Heikki Linnakangas wrote: > On 25.09.2012 10:08, Heikki Linnakangas wrote: > > On 24.09.2012 16:33, Amit Kapila wrote: > >> In any case, it will be better if you can split it into multiple > patches: > >> 1. Having new functionality of "Switching timeline over streaming > >> replication" > >> 2. Refactoring related changes. > Ok, here you go. xlog-c-split-1.patch contains the refactoring of existing code, with no user-visible changes. > streaming-tli-switch-2.patch applies over xlog-c-split-1.patch, and contains the new functionality. > Please find the initial review of the patch. Still more review is pending, > but I thought whatever is done I shall post Some more review: 11. In function readTimeLineHistory() ereport(DEBUG3, (errmsg_internal("history of timeline %u is %s", targetTLI, nodeToString(result)))); Don't nodeToString(result) needs to be changed as it contain list of structure TimeLineHistoryEntry 12. In function @@ -3768,6 +3773,8 @@ rescanLatestTimeLine(void) + * The current timeline must be found in the history file, and the + * next timeline must've forked off from it *after* the current + * recovery location. */ - if (!list_member_int(newExpectedTLIs, - (int) recoveryTargetTLI)) - ereport(LOG, - (errmsg("new timeline %u is not a child of database system timeline %u", - newtarget, - ThisTimeLineID))); is there any logic in the current patch which ensures that above check is not require now? 13. In function @@ -3768,6 +3773,8 @@ rescanLatestTimeLine(void) + found = false; + foreach (cell, newExpectedTLIs) .. .. - list_free(expectedTLIs); + list_free_deep(expectedTLIs); whats the use of the found variable and freeing expectedTLIs in loop might cause problem. 14. In function @@ -3768,6 +3773,8 @@ rescanLatestTimeLine(void) newExpectedTLIs = readTimeLineHistory(newtarget); Shouldn't this variable be declared as newExpectedTLEs as the list returned by readTimeLineHistory contains target list entry 15. StartupXLOG /* Now we can determine the list of expected TLIs */ expectedTLIs = readTimeLineHistory(recoveryTargetTLI); Should expectedTLIs be changed to expectedTLEs as the list returned by readTimeLineHistory contains target list entry 16.@@ -5254,8 +5252,8 @@ StartupXLOG(void) writeTimeLineHistory(ThisTimeLineID, recoveryTargetTLI, - curFileTLI, endLogSegNo, reason); + curFileTLI, EndRecPtr, reason); if endLogSegNo is not used here, it needs to be removd from function declaration as well. 17.@@ -5254,8 +5252,8 @@ StartupXLOG(void) if (InArchiveRecovery) .. .. + + /* The history file can be archived immediately. */ + TLHistoryFileName(histfname, ThisTimeLineID); + XLogArchiveNotify(histfname); Shouldn't this be done archive_mode is on. Right now InArchiveRecovery is true even when we do recovery for standby 18. +static bool +WaitForWALToBecomeAvailable(XLogRecPtr RecPtr, bool randAccess, bool fetching_ckpt) { .. + if (XLByteLT(RecPtr, receivedUpto)) + havedata = true; + else + { + XLogRecPtr latestChunkStart; + + receivedUpto = GetWalRcvWriteRecPtr(&latestChunkStart, &receiveTLI); + if (XLByteLT(RecPtr, receivedUpto) && receiveTLI == curFileTLI) + { + havedata = true; + if (!XLByteLT(RecPtr, latestChunkStart)) + { + XLogReceiptTime = GetCurrentTimestamp(); + SetCurrentChunkStartTime(XLogReceiptTime); + } + } + else + havedata = false; + } In the above logic, it seems there is inconsistency in setting havedata = true; In the above code in else loop, let us say cond. receiveTLI == curFileTLI is false but XLByteLT(RecPtr, receivedUpto) istrue, then in next round in for loop, the check if (XLByteLT(RecPtr, receivedUpto)) will get true and will set havedata = true; which seems to be contradictory. 19. +static bool +WaitForWALToBecomeAvailable(XLogRecPtr RecPtr, bool randAccess, bool fetching_ckpt) { .. + if (PrimaryConnInfo) + { + XLogRecPtr ptr = fetching_ckpt ? RedoStartLSN : RecPtr; + TimeLineID tli = timelineOfPointInHistory(ptr, expectedTLIs); + + if (tli < curFileTLI) I think in some cases if (tli < curFileTLI) might not make sense, as for case where curFileTLI =0 for randAccess. 20. Function name WaitForWALToBecomeAvailable suggests that it waits for WAL, but it also returns true when trigger fileis present, which can be little misleading. 21. @@ -2411,27 +2411,6 @@ reaper(SIGNAL_ARGS) a. won't it impact stop of online basebackup functionality because earlier on promotion I think this code will stopwalsenders and basebackup stop will also give error in such cases. 22. @@ -63,10 +66,17 @@ void_PG_init(void){ /* Tell walreceiver how to reach us */ - if (walrcv_connect != NULL || walrcv_receive != NULL || - walrcv_send != NULL || walrcv_disconnect != NULL) + if (walrcv_connect != NULL || walrcv_identify_system || + walrcv_readtimelinehistoryfile != NULL || check for walrcv_identify_system != NULL is missing. 23. write the function header for newly added functions (libpqrcv_startstreaming, libpqrcv_identify_system, ..) 24. In header of function libpqrcv_receive(), *type needs to be removed.+ * If data was received, returns the length of thedata. *type and *buffer 25. +timeline_history: + K_TIMELINE_HISTORY ICONST + { + TimeLineHistoryCmd *cmd; + + cmd = makeNode(TimeLineHistoryCmd); + cmd->timeline = $2; can handle invalid timeline error same as for opt_timeline 26.@@ -170,6 +187,7 @@ WalReceiverMain(void) + case WALRCV_WAITING: + case WALRCV_STREAMING: /* Shouldn't happen */ elog(PANIC, "walreceiver still running according to shared memory state"); elog message should be changed according to new states. 27.@@ -259,8 +281,11 @@ WalReceiverMain(void) /* Load the libpq-specific functions */ load_file("libpqwalreceiver", false); - if (walrcv_connect == NULL || walrcv_receive == NULL || - walrcv_send == NULL || walrcv_disconnect == NULL) + if (walrcv_connect == NULL || walrcv_startstreaming == NULL || + walrcv_endstreaming == NULL || + walrcv_readtimelinehistoryfile == NULL || + walrcv_receive == NULL || walrcv_send == NULL || + walrcv_disconnect == NULL) check for walrcv_identify_system is missing. 28. +/* + * Check that we're connected to a valid server using the IDENTIFY_SERVER + * replication command, and fetch any timeline history files present in the + * master but missing from this server's pg_xlog directory. + */ +static void +WalRcvHandShake(TimeLineID startpointTLI) In the function header the command name should be IDENTIFY_SYSTEM instead of IDENTIFY_SERVER 29. @@ -170,6 +187,7 @@ WalReceiverMain(void) + * timeline. In case we've already reached the end of the old timeline, + * the server will finish the streaming immediately, and we will + * disconnect. If recovery_target_timeline is 'latest', the startup + * process will then pg_xlog and find the new history file, bump recovery a. I think after reaching at end of old timeline rather than discoonect it will start from new timeline, which will beupdated by startup process. b. The above line seems to be incorrect, "will then (scan/search) pg_xlog" 30. +static void +WalRcvWaitForStartPosition(XLogRecPtr *startpoint, TimeLineID *startpointTLI) /* + * nudge startup process to notice that we've stopped streaming and are + * now waiting for instructions. + */ + WakeupRecovery(); for (;;) { In this for loop don't we need to check interrupts or postmaster alive or recovery in progress so that if any other process continues, it should not wait indefinately. 31.+static void +WalRcvWaitForStartPosition(XLogRecPtr *startpoint, TimeLineID *startpointTLI) /* + * nudge startup process to notice that we've stopped streaming and are + * now waiting for instructions. + */ + WakeupRecovery(); for (;;) { + SpinLockAcquire(&walrcv->mutex); + if (walrcv->walRcvState == WALRCV_STARTING) { I think it can reach WALRCV_STOPPING state also after WALRCV_WAITING from shutdown, so we should check for that state as well. 32.@@ -170,6 +187,7 @@ WalReceiverMain(void) { .. .. + elog(LOG, "walreceiver ended streaming and awaits new instructions"); + WalRcvWaitForStartPosition(&startpoint, &startpointTLI); a. After getting new startpoint and tli, it will go again for WalRcvHandShake(startpointTLI); so here chances are there,it will again fetch the history files from server which we have already fetched. b. Also Identify_system command will run again and get the information such as system identifier, which is completely redundantat this point. It fetches tli from primary also which I think can't be changed from what earlier it has fetched. 33. .@@ -170,6 +187,7 @@ WalReceiverMain(void) + for (;;) + { + if (len > 0) + XLogWalRcvProcessMsg(buf[0], &buf[1], len - 1); + else if (len == 0) + break; + else if (len < 0) + { + ereport(LOG, + (errmsg("replication terminated by primary server"), + errdetail("End of WAL reached on timeline %u", startpointTLI))); + endofwal = true; + break; + } + len = walrcv_receive(0, &buf); + } + + /* Let the master know that we received some data. */ + XLogWalRcvSendReply(); + + /* + * If we've written some records, flush them to disk and let the + * startup process and primary server know about them. + */ + XLogWalRcvFlush(false); a. In the above code in for loop, when it breaks due to len < 0, there is no need to send reply to master. b. also when it breaks due to len < 0, there can be 2 reasons, one is end of copy mode or primary server has disconnected.I think in second case handling should be same as what without this feature. Not sure if its eventually turningout to be same. 34. +bool +WalRcvStreaming(void) +{ In this function, right now if state is WALRCV_WAITING, then it will return false. I think waiting is better than starting for the matter of checking if walreceiver is in progress. or is state WALRCV_WAITING anytime expected when this function is called, if not then we log the error for invalid state. With Regards, Amit Kapila.
Thanks for the thorough review! I committed the xlog.c refactoring patch now. Attached is a new version of the main patch, comments on specific points below. I didn't adjust the docs per your comments yet, will do that next. On 01.10.2012 05:25, Amit kapila wrote: > 1. In function readTimeLineHistory(), > two mechanisms are used to fetch timeline from history file > + sscanf(fline, "%u\t%X/%X", &tli, &switchpoint_hi, > &switchpoint_lo); > + > + /* expect a numeric timeline ID as first field of line */ > + tli = (TimeLineID) strtoul(ptr, &endptr, 0); > If we use new mechanism, it will not be able to detect error as it is > doing in current case. Fixed, by checking the return value of sscanf(). > 2. In function readTimeLineHistory(), > + fd = AllocateFile(path, "r"); > + if (fd == NULL) > + { > + if (errno != ENOENT) > + ereport(FATAL, > + (errcode_for_file_access(), > + errmsg("could not open file > \"%s\": %m", path))); > + /* Not there, so assume no parents */ > + return list_make1_int((int) targetTLI); > + } > still return list_make1_int((int) targetTLI); is used. Fixed. > 3. Function timelineOfPointInHistory(), should return the timeline of recptr > passed to it. > a. is it okay to decide based on xlog recordpointer that which timeline > it belongs to, as different > timelines can have same xlog recordpointer? In a particular timeline, the history is linear, and a given point in WAL unambiguously has one timeline ID. There might be some other timelines that branch off at different points, but once you pick a particular timeline, you can unambiguously trace it all the way to the beginning of WAL, and tell what the timeline ID of each point in WAL was. > b. it seems from logic that it will return timeline previous to the > timeline of recptr passed. > For example if the timeline 3's switchpoint is equal to recptr passed > then it will return timeline 2. I expanded the comment in the function a bit, I hope it makes more sense now. > 4. In writeTimeLineHistory function variable endTLI is never used. Removed. > 5. In header of function writeTimeLineHistory(), can give explanation about > XLogRecPtr switchpoint Added. > 6. @@ -6869,11 +5947,35 @@ StartupXLOG(void) > */ > if (InArchiveRecovery) > { > + char reason[200]; > + > > > + /* > + * Write comment to history file to explain why and where > timeline > + * changed. Comment varies according to the recovery target > used. > + */ > + if (recoveryTarget == RECOVERY_TARGET_XID) > + snprintf(reason, sizeof(reason), > + "%s transaction %u", > + recoveryStopAfter ? "after" : > "before", > + recoveryStopXid); > > In the comment above this line you mentioned why and where timeline changed. > > However in the reason field only specifies about where part. I didn't change this in the patch. I guess it's not obvious, but you can deduce the 'why' part from the message. > 7. + * Returns the redo pointer of the "previous" checkpoint. > +GetOldestRestartPoint(XLogRecPtr *oldrecptr, TimeLineID *oldtli) > +{ > + if (InRedo) > + { > + LWLockAcquire(ControlFileLock, LW_SHARED); > + *oldrecptr = ControlFile->checkPointCopy.redo; > + *oldtli = ControlFile->checkPointCopy.ThisTimeLineID; > + LWLockRelease(ControlFileLock); > + } > > a. In this function, is it required to take ControlFileLock as earlier also > there was no lock to protect this read > when it get called from RestoreArchivedFile, and I think at this point no > one else can modify these values. > However for code consistency purpose like whenever or wherever read the > controlfile values, read it with read lock. Yeah, it's just for the sake of consistency. > b. As per your comment it should have returned "previous" checkpoint, > however the function returns values of > latest checkpoint. Changed the comment. I wonder if we should be more conservative, and really keep the WAL back to the "previous" checkpoint, but I won't change that as part of this patch. > 8. In function writeTimeLineHistoryFile(), will it not be better to directly > write rather than to later do pg_fsync(). > as it is just one time write. Not sure I understood this right, but writeTimeLineHistoryFile() needs to avoid putting a corrupt, e.g incomplete, file in pg_xlog. The same as writeTimeLineHistory(). That's why the write+fsync+rename dance is needed. > 9. +XLogRecPtr > +timeLineSwitchPoint(XLogRecPtr startpoint, TimeLineID tli) > .. > .. > + * starting point. This is because the client can > legimately > spelling of legitimately needs to be corrected. Fixed. > 10.+XLogRecPtr > +timeLineSwitchPoint(XLogRecPtr startpoint, TimeLineID tli) > .. > .. > + if (tli < ThisTimeLineID) > + { > + if (!nexttle) > + elog(ERROR, "could not find history entry for child > of timeline %u", tli); /* shouldn't happen */ > + } > > I don't understand the meaning of the above check, as I think this > situation can occur when this function gets called from StartReplication, > because always tli sent by standby to new master will be less than > ThisTimeLineID and it can be first in list. Note that the list is in newest-first order. Ie. the last line in the history file is first in the list. The first entry in the list is always for ThisTimeLineID, which is why the (tli < ThisTimeLineID && !nexttle) combination isn't possible. > 11. In function readTimeLineHistory() > ereport(DEBUG3, > (errmsg_internal("history of timeline %u is %s", > targetTLI, nodeToString(result)))); > > > Don't nodeToString(result) needs to be changed as it contain list of structure TimeLineHistoryEntry Yep. Since this is just a DEBUG3, I'll just remove that, rather than add the extra code needed to keep the output. > 12. In function @@ -3768,6 +3773,8 @@ rescanLatestTimeLine(void) > + * The current timeline must be found in the history file, and the > + * next timeline must've forked off from it *after* the current > + * recovery location. > */ > - if (!list_member_int(newExpectedTLIs, > - (int) recoveryTargetTLI)) > - ereport(LOG, > - (errmsg("new timeline %u is not a child of database system timeline %u", > - newtarget, > - ThisTimeLineID))); > > > is there any logic in the current patch which ensures that above check is not require now? > > > 13. In function @@ -3768,6 +3773,8 @@ rescanLatestTimeLine(void) > + found = false; > + foreach (cell, newExpectedTLIs) > .. > .. > - list_free(expectedTLIs); > + list_free_deep(expectedTLIs); > whats the use of the found variable and freeing expectedTLIs in loop might cause problem. Oops, there's some code missing there. Apparently I botched that at some point while splitting the patch into two. Fixed. > 14. In function @@ -3768,6 +3773,8 @@ rescanLatestTimeLine(void) > newExpectedTLIs = readTimeLineHistory(newtarget); > Shouldn't this variable be declared as newExpectedTLEs as the list returned by readTimeLineHistory contains target listentry > > > 15. StartupXLOG > /* Now we can determine the list of expected TLIs */ > expectedTLIs = readTimeLineHistory(recoveryTargetTLI); > > > Should expectedTLIs be changed to expectedTLEs as the list returned by readTimeLineHistory contains target list entry Makes sense, renamed these two. > 16.@@ -5254,8 +5252,8 @@ StartupXLOG(void) > writeTimeLineHistory(ThisTimeLineID, recoveryTargetTLI, > - curFileTLI, endLogSegNo, reason); > + curFileTLI, EndRecPtr, reason); > if endLogSegNo is not used here, it needs to be removd from function declaration as well. I didn't understand this one. endLogSegNo is still used earlier in StartupXLOG. > 17.@@ -5254,8 +5252,8 @@ StartupXLOG(void) > if (InArchiveRecovery) > .. > .. > + > + /* The history file can be archived immediately. */ > + TLHistoryFileName(histfname, ThisTimeLineID); > + XLogArchiveNotify(histfname); > > > Shouldn't this be done archive_mode is on. Right now InArchiveRecovery is true even when we do recovery for standby Fixed. > 18. +static bool > +WaitForWALToBecomeAvailable(XLogRecPtr RecPtr, bool randAccess, bool fetching_ckpt) > { > .. > + if (XLByteLT(RecPtr, receivedUpto)) > + havedata = true; > + else > + { > + XLogRecPtr latestChunkStart; > + > + receivedUpto = GetWalRcvWriteRecPtr(&latestChunkStart,&receiveTLI); > + if (XLByteLT(RecPtr, receivedUpto)&& receiveTLI == curFileTLI) > + { > + havedata = true; > + if (!XLByteLT(RecPtr, latestChunkStart)) > + { > + XLogReceiptTime = GetCurrentTimestamp(); > + SetCurrentChunkStartTime(XLogReceiptTime); > + } > + } > + else > + havedata = false; > + } > > > In the above logic, it seems there is inconsistency in setting havedata = true; > In the above code in else loop, let us say cond. receiveTLI == curFileTLI is false but XLByteLT(RecPtr, receivedUpto) istrue, > then in next round in for loop, the check if (XLByteLT(RecPtr, receivedUpto)) will get true and will set havedata = true; > which seems to be contradictory. Hmm, I think you're saying that we should check that receiveTLI == curFileTLI also in the first if-statement above. Did that. > 19. > > > +static bool > +WaitForWALToBecomeAvailable(XLogRecPtr RecPtr, bool randAccess, bool fetching_ckpt) > { > .. > + if (PrimaryConnInfo) > + { > + XLogRecPtr ptr = fetching_ckpt ? RedoStartLSN : RecPtr; > + TimeLineID tli = timelineOfPointInHistory(ptr, expectedTLIs); > + > + if (tli< curFileTLI) > > > I think in some cases if (tli< curFileTLI) might not make sense, as for case where curFileTLI =0 for randAccess. Well, if curFileTLI == 0, then it's surely < tli. I think that's the correct behavior, but I'll add an explicit check for randAccess to make it more explicit. > 20. Function name WaitForWALToBecomeAvailable suggests that it waits for WAL, but it also returns true when trigger fileis present, > which can be little misleading. Added a comment above the function to clarify that. > 21. @@ -2411,27 +2411,6 @@ reaper(SIGNAL_ARGS) > > a. won't it impact stop of online basebackup functionality because earlier on promotion > I think this code will stop walsenders and basebackup stop will also give error in such cases. Hmm, should a base backup be aborted when the standby is promoted? Does the promotion render the backup corrupt? > 22. @@ -63,10 +66,17 @@ void > _PG_init(void) > { > /* Tell walreceiver how to reach us */ > - if (walrcv_connect != NULL || walrcv_receive != NULL || > - walrcv_send != NULL || walrcv_disconnect != NULL) > + if (walrcv_connect != NULL || walrcv_identify_system || > + walrcv_readtimelinehistoryfile != NULL || > > > check for walrcv_identify_system != NULL is missing. Fixed. > 23. write the function header for newly added functions (libpqrcv_startstreaming, libpqrcv_identify_system, ..) Fixed. > 24. In header of function libpqrcv_receive(), *type needs to be removed. > + * If data was received, returns the length of the data. *type and *buffer Fixed. > 25. > +timeline_history: > + K_TIMELINE_HISTORY ICONST > + { > + TimeLineHistoryCmd *cmd; > + > + cmd = makeNode(TimeLineHistoryCmd); > + cmd->timeline = $2; > > > can handle invalid timeline error same as for opt_timeline Fixed. > 26.@@ -170,6 +187,7 @@ WalReceiverMain(void) > + case WALRCV_WAITING: > + case WALRCV_STREAMING: > /* Shouldn't happen */ > elog(PANIC, "walreceiver still running according to shared memory state"); > elog message should be changed according to new states. I think it's ok as is. Both 'waiting' and 'streaming' can be thought of as 'running'. WalRcvRunning() also returns true for both states. > 27.@@ -259,8 +281,11 @@ WalReceiverMain(void) > > /* Load the libpq-specific functions */ > load_file("libpqwalreceiver", false); > - if (walrcv_connect == NULL || walrcv_receive == NULL || > - walrcv_send == NULL || walrcv_disconnect == NULL) > + if (walrcv_connect == NULL || walrcv_startstreaming == NULL || > + walrcv_endstreaming == NULL || > + walrcv_readtimelinehistoryfile == NULL || > + walrcv_receive == NULL || walrcv_send == NULL || > + walrcv_disconnect == NULL) > > > check for walrcv_identify_system is missing. Fixed. > 28. > +/* > + * Check that we're connected to a valid server using the IDENTIFY_SERVER > + * replication command, and fetch any timeline history files present in the > + * master but missing from this server's pg_xlog directory. > + */ > +static void > +WalRcvHandShake(TimeLineID startpointTLI) > > > In the function header the command name should be IDENTIFY_SYSTEM instead of IDENTIFY_SERVER Fixed. > 29. @@ -170,6 +187,7 @@ WalReceiverMain(void) > + * timeline. In case we've already reached the end of the old timeline, > + * the server will finish the streaming immediately, and we will > + * disconnect. If recovery_target_timeline is 'latest', the startup > + * process will then pg_xlog and find the new history file, bump recovery > > > a. I think after reaching at end of old timeline rather than discoonect it will start from new timeline, to> which will be updated by startup process. > b. The above line seems to be incorrect, "will then (scan/search) pg_xlog" Fixed the comment. > 30. +static void > +WalRcvWaitForStartPosition(XLogRecPtr *startpoint, TimeLineID *startpointTLI) > > > /* > + * nudge startup process to notice that we've stopped streaming and are > + * now waiting for instructions. > + */ > + WakeupRecovery(); > for (;;) > { > > > In this for loop don't we need to check interrupts or postmaster alive or recovery in progress > so that if any other process continues, it should not wait indefinately. Added a ProcessWalRcvInterrupts() check. There is a PostmasterIsAlive() check there already. > 31.+static void > +WalRcvWaitForStartPosition(XLogRecPtr *startpoint, TimeLineID *startpointTLI) > > > /* > + * nudge startup process to notice that we've stopped streaming and are > + * now waiting for instructions. > + */ > + WakeupRecovery(); > for (;;) > { > > > + SpinLockAcquire(&walrcv->mutex); > + if (walrcv->walRcvState == WALRCV_STARTING) > { > > > I think it can reach WALRCV_STOPPING state also after WALRCV_WAITING from shutdown, > so we should check for that state as well. Added that, although we should receive the SIGTERM when the startup process wants the walreceiver to die. > 32.@@ -170,6 +187,7 @@ WalReceiverMain(void) > { > .. > .. > > > + elog(LOG, "walreceiver ended streaming and awaits new instructions"); > + WalRcvWaitForStartPosition(&startpoint,&startpointTLI); > > > a. After getting new startpoint and tli, it will go again for WalRcvHandShake(startpointTLI); > so here chances are there, it will again fetch the history files from server which we have > already fetched. WalRcvHandShake() only fetches history files that don't exist in pg_xlog already. > b. Also Identify_system command will run again and get the information such as system identifier, > which is completely redundant at this point. > It fetches tli from primary also which I think can't be changed from what earlier it has fetched. The server's tli might've changed in cascading replication, where the server is also a standby running with recovery_target_timeline='latest'. It's fairly unlikely if we just ran the Identify_system command, but I'm inclined to keep that. One extra roundtrip isn't that bad, and I think it'd complicate the logic to try to avoid that. > 33. .@@ -170,6 +187,7 @@ WalReceiverMain(void) > + for (;;) > + { > + if (len> 0) > + XLogWalRcvProcessMsg(buf[0],&buf[1], len - 1); > + else if (len == 0) > + break; > + else if (len< 0) > + { > + ereport(LOG, > + (errmsg("replication terminated by primary server"), > + errdetail("End of WAL reached on timeline %u", startpointTLI))); > + endofwal = true; > + break; > + } > + len = walrcv_receive(0,&buf); > + } > + > + /* Let the master know that we received some data. */ > + XLogWalRcvSendReply(); > + > + /* > + * If we've written some records, flush them to disk and let the > + * startup process and primary server know about them. > + */ > + XLogWalRcvFlush(false); > > > a. In the above code in for loop, when it breaks due to len< 0, there is no need to send reply to master. Well, I think it's prudent to send one more reply at the end of streaming. > b. also when it breaks due to len< 0, there can be 2 reasons, one is end of copy mode or primary server has > disconnected. I think in second case handling should be same as what without this feature. > Not sure if its eventually turning out to be same. No, libpq_receive() will throw an error if the connection is broken for some reason. It only returns -1 at end-of-copy. > 34. > +bool > +WalRcvStreaming(void) > +{ > > > In this function, right now if state is WALRCV_WAITING, then it will return false. > I think waiting is better than starting for the matter of checking if walreceiver is in progress. > or is state WALRCV_WAITING anytime expected when this function is called, if not then we log the > error for invalid state. It's normal to call WalRcvStreaming() when it's in waiting mode. WalRcvStreaming() is called from WaitForWALToBecomeAvailable, always, regardless of whether streaming replication is even enabled. Thanks again for the detailed review! - Heikki
Attachment
On Tuesday, October 02, 2012 4:21 PM Heikki Linnakangas wrote: > Thanks for the thorough review! I committed the xlog.c refactoring patch > now. Attached is a new version of the main patch, comments on specific > points below. I didn't adjust the docs per your comments yet, will do > that next. I have some doubts regarding the comments fixed by you and some more new review comments. After this I shall focus majorly towards testing of this Patch. > On 01.10.2012 05:25, Amit kapila wrote: > > 1. In function readTimeLineHistory(), > > two mechanisms are used to fetch timeline from history file > > + sscanf(fline, "%u\t%X/%X", &tli, &switchpoint_hi, > > &switchpoint_lo); > > + > > > 8. In function writeTimeLineHistoryFile(), will it not be better to > > directly write rather than to later do pg_fsync(). > > as it is just one time write. > > Not sure I understood this right, but writeTimeLineHistoryFile() needs > to avoid putting a corrupt, e.g incomplete, file in pg_xlog. The same as > writeTimeLineHistory(). That's why the write+fsync+rename dance is > needed. > Why fsync, isn't the above purpose be resolved if write directly writes to file and then rename. > > 21. @@ -2411,27 +2411,6 @@ reaper(SIGNAL_ARGS) > > > > a. won't it impact stop of online basebackup functionality > because earlier on promotion > > I think this code will stop walsenders and basebackup stop > will also give error in such cases. > > Hmm, should a base backup be aborted when the standby is promoted? Does > the promotion render the backup corrupt? I think currently it does so. Pls refer 1. do_pg_stop_backup(char *labelfile, bool waitforarchive) { .. if (strcmp(backupfrom, "standby") == 0 && !backup_started_in_recovery) ereport(ERROR, (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), errmsg("the standby was promoted during online backup"), errhint("This means that the backup being taken is corrupt " "and should not be used. " "Try taking another online backup."))); .. } 2. In documentation of pg_basebackup there is a Note: .If the standby is promoted to the master during online backup, the backup fails. New Ones --------------- 35.WalSenderMain(void) { .. + if (walsender_shutdown_requested) + ereport(FATAL, + (errcode(ERRCODE_ADMIN_SHUTDOWN), + errmsg("terminating replication connection due to administrator command"))); + + /* Tell the client that we are ready to receive commands */ + ReadyForQuery(DestRemote); + .. + if (walsender_shutdown_requested) + ereport(FATAL, + (errcode(ERRCODE_ADMIN_SHUTDOWN), + errmsg("terminating replication connection due to administrator command"))); + is it necessary to check walsender_shutdown_requested 2 times in a loop, if yes, then can we write comment why it is important to check it again. 35. +SendTimeLineHistory(TimeLineHistoryCmd *cmd) { .. + fd = PathNameOpenFile(path, O_RDONLY | PG_BINARY, 0666); errorhandling for fd < 0 is missing. 36.+SendTimeLineHistory(TimeLineHistoryCmd *cmd) { .. if (nread <= 0) + ereport(ERROR, + (errcode_for_file_access(), + errmsg("could not read file \"%s\": %m", + path))); FileClose should be done in error case as well. 37. static void XLogSend(char *msgbuf, bool *caughtup) { .. if (currentTimeLineIsHistoric && XLByteLE(currentTimeLineValidUpto, sentPtr)) { /* * This was a historic timeline, and we've reached the point where * we switched to the next timeline. Let the client know we've * reached the end of this timeline, and what the next timeline is. */ /* close the current file. */ if (sendFile >= 0) close(sendFile); sendFile = -1; *caughtup = true; /* Send CopyDone */ pq_putmessage_noblock('c', NULL, 0); streamingDoneSending = true; return; } } I am not able to understand after sending CopyDone message from above code, how walreceiver is handling it and then replying it a CopyDone message. Basically I want to know the walreceiver code which handles it? 38. static void WalSndLoop(void) { @@ -756,18 +898,24 @@ WalSndLoop(void) /* Normal exit from the walsender is here */ if (walsender_shutdown_requested) - { - /* Inform the standby that XLOG streaming is done */ - pq_puttextmessage('C', "COPY 0"); - pq_flush(); - - proc_exit(0); - } + ereport(FATAL, + (errcode(ERRCODE_ADMIN_SHUTDOWN), + errmsg("terminating replication connection due to administrator command"))); What is the reason of removal of sending above message to standby when shutdown was requested? 39. WalSndLoop(void) { .. /* If nothing remains to be sent right now ... */ if (caughtup && !pq_is_send_pending()) { /* * If we're in catchup state, move to streaming. This is an * important state change for users to know about, since before * this point data loss might occur if the primary dies and we * need to failover to the standby. The state change is also * important for synchronous replication, since commits that * started to wait at that point might wait for some time. */ if (MyWalSnd->state == WALSNDSTATE_CATCHUP) { ereport(DEBUG1, (errmsg("standby \"%s\" has now caught up with primary", application_name))); WalSndSetState(WALSNDSTATE_STREAMING); } .. } After new implementation, I think above if loop [if (caughtup && !pq_is_send_pending())] can be true when the standby has not actually caught up as now it sends the data from previous timelines. With Regards, Amit Kapila.
Promoting a standby during base backup (was Re: Switching timeline over streaming replication)
From
Heikki Linnakangas
Date:
On 03.10.2012 18:15, Amit Kapila wrote: > On Tuesday, October 02, 2012 4:21 PM Heikki Linnakangas wrote: >> Hmm, should a base backup be aborted when the standby is promoted? Does >> the promotion render the backup corrupt? > > I think currently it does so. Pls refer > 1. > do_pg_stop_backup(char *labelfile, bool waitforarchive) > { > .. > if (strcmp(backupfrom, "standby") == 0&& !backup_started_in_recovery) > ereport(ERROR, > > (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), > errmsg("the standby was promoted during > online backup"), > errhint("This means that the backup being > taken is corrupt " > "and should not be used. " > "Try taking another online > backup."))); > .. > > } Okay. I think that check in do_pg_stop_backup() actually already ensures that you don't end up with a corrupt backup, even if the standby is promoted while a backup is being taken. Admittedly it would be nicer to abort it immediately rather than error out at the end. But I wonder why promoting a standby renders the backup invalid in the first place? Fujii, Simon, can you explain that? - Heikki
> On Wednesday, October 03, 2012 8:45 PM Heikki Linnakangas wrote: > On Tuesday, October 02, 2012 4:21 PM Heikki Linnakangas wrote: > > Thanks for the thorough review! I committed the xlog.c refactoring > patch > > now. Attached is a new version of the main patch, comments on specific > > points below. I didn't adjust the docs per your comments yet, will do > > that next. > > I have some doubts regarding the comments fixed by you and some more new > review comments. > After this I shall focus majorly towards testing of this Patch. > Testing ----------- Failed Case -------------- 1. promotion of standby to master and follow standby to new master. 2. Stop standby and master. Restart standby first and then master 3. Restart of standby gives below errors E:\pg_git_code\installation\bin>LOG: database system was shut down in recovery at 2012-10-04 18:36:00 IST LOG: entering standby mode LOG: consistent recovery state reached at 0/176B800 LOG: redo starts at 0/176B800 LOG: record with zero length at 0/176BD68 LOG: database system is ready to accept read only connections LOG: streaming replication successfully connected to primary LOG: out-of-sequence timeline ID 1 (after 2) in log segment 0000000200000000000 00001, offset 0 FATAL: terminating walreceiver process due to administrator command LOG: out-of-sequence timeline ID 1 (after 2) in log segment 0000000200000000000 00001, offset 0 LOG: out-of-sequence timeline ID 1 (after 2) in log segment 0000000200000000000 00001, offset 0 LOG: out-of-sequence timeline ID 1 (after 2) in log segment 0000000200000000000 00001, offset 0 LOG: out-of-sequence timeline ID 1 (after 2) in log segment 0000000200000000000 00001, offset 0 Once this error comes, restart master/standby in any order or do some operations on master, always there is above error On standby. Passed Cases ------------- 1. After promoting standby as new master, try to make previous master (having same WAL as new master) as standby. In this case recovery.conf recovery_target_timeline set to latest. It ables to connect to new master and started streaming as per expectation. - As per expected behavior. 2. After promoting standby as new master, try to make previous master (having more WAL compare to new master) as standby, error is displayed. - As per expected behavior 3. After promoting standby as new master, try to make previous master (having same WAL as new master) as standby. In this case recovery.conf recovery_target_timeline is not set. Following LOG is displayed. LOG: fetching timeline history file for timeline 2 from primary server LOG: replication terminatedby primary server DETAIL: End of WAL reached on timeline 1 LOG: walreceiver ended streaming and awaits newinstructions LOG: re-handshaking at position 0/1000000 on tli 1 LOG: replication terminated by primary server DETAIL: End of WAL reached on timeline 1 LOG: walreceiver ended streaming and awaits new instructions LOG: re-handshakingat position 0/1000000 on tli 1 LOG: replication terminated by primary server DETAIL: End of WAL reachedon timeline 1 - As per expected behavior Pending Cases which needs to be tested (these are scenarios, some more testing I will do based on these scenarios) --------------------------------------- 1. a. Master M-1 b. Standby S-1 follows M-1 c. Standby S-2 follows M-1 d. Promote S-1 as master e. Try to followS-2 to S-1 -- operation should be success 2. a. Master M-1 b. Standby S-1 follows M-1 c. Stop S-1, M-1 d. Do the PITR in M-1 2 times. This is to increment timelinein M-1 e. try to follow standby S-1 to M-1 -- it should be success. 3. a. Master M-1 b. Standby S-1, S-2 follows M1 c. Standby S-3, S-4 follows S-1 d. Promote Standby which has highestWAL. e. follow all standby's to the new master. 4. a. Master M-1 b. Synchronous Standby S-1, S-2 c. Promote S-1 d. Follow M-1, S-2 to S-1 -- this operation shouldbe success. Concurrent Operations --------------------------- 1. a. Master M-1 , Standby S-1 follows M-1, Standby S-2 follows M-1 b. Many concurrent operations on master M-1 c. During concurrent ops, Promote S-1 d. try S-2 to followS-1 -- it should happen successfully. 2. During Promotion, call pg_basebackup 3. During Promotion, try to connect client Resource Testing ------------------ 1. a.Make standby follow master which is many time lines ahead b. Observeif there is any resource leak c. Allow the streaming replication for 30 mins d. Observe if there is anyresource leak Code Review ------------- Libpqrcv_readtimelinehistoryfile() { .. .. + if (PQnfields(res) != 2 || PQntuples(res) != 1) + { + int ntuples = PQntuples(res); + int nfields = PQnfields(res); + + PQclear(res); + ereport(ERROR, + (errmsg("invalid response from primary server"), + errdetail("Expected 1 tuple with 3 fields, got %d tuples with %d fields.", + ntuples, nfields))); + } .. } The error message is saying 3 fields needs to be read in timeline history, but the check seems to be is done for 2 fields. Kindly let me know if you want me to focus on any other areas for testing this feature. With Regards, Amit Kapila.
Sharing more infrastructure between walsenders and regular backends (was Re: Switching timeline over streaming replication)
From
Heikki Linnakangas
Date:
On 03.10.2012 18:15, Amit Kapila wrote: > 35.WalSenderMain(void) > { > .. > + if (walsender_shutdown_requested) > + ereport(FATAL, > + (errcode(ERRCODE_ADMIN_SHUTDOWN), > + errmsg("terminating replication > connection due to administrator command"))); > + > + /* Tell the client that we are ready to receive commands */ > > + ReadyForQuery(DestRemote); > + > .. > > + if (walsender_shutdown_requested) > + ereport(FATAL, > + (errcode(ERRCODE_ADMIN_SHUTDOWN), > + errmsg("terminating replication > connection due to administrator command"))); > + > > is it necessary to check walsender_shutdown_requested 2 times in a > loop, if yes, then can we write comment why it is important to check > it again. The idea was to check for shutdown request before and after the pq_getbyte() call, because that can block for a long time. Looking closer, we don't currently (ie. without this patch) make any effort to react to SIGTERM in a walsender, while it's waiting for a command from the client. After starting replication, it does check walsender_shutdown_requested in the loop, and it's also checked during a base backup (although only when switching to send next file, which seems too seldom). This issue is orthogonal to handling timeline changes over streaming replication, although that patch will make it more important to handle SIGTERM quickly while waiting for a command, because you stay in that mode for longer and more often. I think walsender needs to share more infrastructure with regular backends to handle this better. When we first implemented streaming replication in 9.0, it made sense to implement just the bare minimum needed to accept the handshake commands before entering the Copy state, but now that the replication command set has grown to cover base backups, and fetching timelines with the patch being discussed, we should bite the bullet and make the command loop more feature-complete and robust. In a regular backend, the command loop reacts to SIGTERM immediately, setting ImmediateInterruptOK at the right places, and calling CHECK_FOR_INTERRUPTS() at strategic places. I propose that we let PostgresMain handle the main command loop for walsender processes too, like it does for regular backends, and use ProcDiePending and the regular die() signal handler for SIGTERM in walsender as well. So I propose the attached patch. I made small changes to postgres.c to make it call exec_replication_command() instead of exec_simple_query(), and reject extend query protocol, in a WAL sender process. A lot of code related to handling the main command loop and signals is removed from walsender.c. - Heikki
Attachment
Re: Sharing more infrastructure between walsenders and regular backends (was Re: Switching timeline over streaming replication)
From
Tom Lane
Date:
Heikki Linnakangas <hlinnakangas@vmware.com> writes: > So I propose the attached patch. I made small changes to postgres.c to > make it call exec_replication_command() instead of exec_simple_query(), > and reject extend query protocol, in a WAL sender process. A lot of code > related to handling the main command loop and signals is removed from > walsender.c. Why do we need the forbidden_in_wal_sender stuff? If we're going in this direction, I suggest there is little reason to restrict what the replication client can do. This seems to be both ugly and a drag on the performance of normal backends. regards, tom lane
Re: Sharing more infrastructure between walsenders and regular backends (was Re: Switching timeline over streaming replication)
From
Heikki Linnakangas
Date:
On 04.10.2012 19:00, Tom Lane wrote: > Heikki Linnakangas<hlinnakangas@vmware.com> writes: >> So I propose the attached patch. I made small changes to postgres.c to >> make it call exec_replication_command() instead of exec_simple_query(), >> and reject extend query protocol, in a WAL sender process. A lot of code >> related to handling the main command loop and signals is removed from >> walsender.c. > > Why do we need the forbidden_in_wal_sender stuff? If we're going in > this direction, I suggest there is little reason to restrict what the > replication client can do. This seems to be both ugly and a drag on > the performance of normal backends. Well, there's not much need for parameterized queries or cursors with the replication command set at the moment. I don't think it's worth it to try to support them. Fastpath function calls make no sense either, as you can't call user-defined functions in a walsender anyway. Perhaps we could make walsenders even more like regular backends than what I was proposing, so that the replication commands are parsed and executed just like regular utility commands. However, that'd require some transaction support in walsender, for starters, which seems messy. It might become sensible in the future if the replication command set gets even more complicated, but it doesn't seem like a good idea at the moment. - Heikki
Re: Promoting a standby during base backup (was Re: Switching timeline over streaming replication)
From
Fujii Masao
Date:
On Thu, Oct 4, 2012 at 4:59 PM, Heikki Linnakangas <hlinnakangas@vmware.com> wrote: > On 03.10.2012 18:15, Amit Kapila wrote: >> >> On Tuesday, October 02, 2012 4:21 PM Heikki Linnakangas wrote: >>> >>> Hmm, should a base backup be aborted when the standby is promoted? Does >>> the promotion render the backup corrupt? >> >> >> I think currently it does so. Pls refer >> 1. >> do_pg_stop_backup(char *labelfile, bool waitforarchive) >> { >> .. >> if (strcmp(backupfrom, "standby") == 0&& !backup_started_in_recovery) >> ereport(ERROR, >> >> (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), >> errmsg("the standby was promoted during >> online backup"), >> errhint("This means that the backup >> being >> taken is corrupt " >> "and should not be used. >> " >> "Try taking another >> online >> backup."))); >> .. >> >> } > > > Okay. I think that check in do_pg_stop_backup() actually already ensures > that you don't end up with a corrupt backup, even if the standby is promoted > while a backup is being taken. Admittedly it would be nicer to abort it > immediately rather than error out at the end. > > But I wonder why promoting a standby renders the backup invalid in the first > place? Fujii, Simon, can you explain that? Simon had the same question and I answered it before. http://archives.postgresql.org/message-id/CAHGQGwFU04oO8YL5SUcdjVq3BRNi7WtfzTy9wA2kXtZNHicTeA@mail.gmail.com --------------------------------------- > You say > "If the standby is promoted to the master during online backup, the > backup fails." > but no explanation of why? > > I could work those things out, but I don't want to have to, plus we > may disagree if I did. If the backup succeeds in that case, when we start an archive recovery from that backup, the recovery needs to cross between two timelines. Which means that we need to set recovery_target_timeline before starting recovery. Whether recovery_target_timeline needs to be set or not depends on whether the standby was promoted during taking the backup. Leaving such a decision to a user seems fragile. pg_basebackup -x ensures that all required files are included in the backup and we can start recovery without restoring any file from the archive. But if the standby is promoted during the backup, the timeline history file would become an essential file for recovery, but it's not included in the backup. --------------------------------------- The situation may change if your switching-timeline patch has been committed. It's useful if we can continue the backup even if the standby is promoted. Regards, -- Fujii Masao
Re: Sharing more infrastructure between walsenders and regular backends (was Re: Switching timeline over streaming replication)
From
Simon Riggs
Date:
On 4 October 2012 17:23, Heikki Linnakangas <hlinnakangas@vmware.com> wrote: > On 04.10.2012 19:00, Tom Lane wrote: >> >> Heikki Linnakangas<hlinnakangas@vmware.com> writes: >>> >>> So I propose the attached patch. I made small changes to postgres.c to >>> make it call exec_replication_command() instead of exec_simple_query(), >>> and reject extend query protocol, in a WAL sender process. A lot of code >>> related to handling the main command loop and signals is removed from >>> walsender.c. >> >> >> Why do we need the forbidden_in_wal_sender stuff? If we're going in >> this direction, I suggest there is little reason to restrict what the >> replication client can do. This seems to be both ugly and a drag on >> the performance of normal backends. > > > Well, there's not much need for parameterized queries or cursors with the > replication command set at the moment. I don't think it's worth it to try to > support them. Fastpath function calls make no sense either, as you can't > call user-defined functions in a walsender anyway. > > Perhaps we could make walsenders even more like regular backends than what I > was proposing, so that the replication commands are parsed and executed just > like regular utility commands. However, that'd require some transaction > support in walsender, for starters, which seems messy. It might become > sensible in the future if the replication command set gets even more > complicated, but it doesn't seem like a good idea at the moment. It's come up a few times now that people want to run a few queries either before or after running a base backup. Since the pg_basebackup stuff uses walsender, this make such things impossible. So to support that, we need to allow two kinds of connection, one to "replication" and one to something else, and since the something else is not guaranteed to exist that makes it even harder. Andres suggested to me the other day we make walsender more like regular backends. At the time I wasn't sure I agreed, but reading this it looks like a sensible way to go. -- Simon Riggs http://www.2ndQuadrant.com/PostgreSQL Development, 24x7 Support, Training & Services
Re: Sharing more infrastructure between walsenders and regular backends (was Re: Switching timeline over streaming replication)
From
Tom Lane
Date:
Simon Riggs <simon@2ndQuadrant.com> writes: > On 4 October 2012 17:23, Heikki Linnakangas <hlinnakangas@vmware.com> wrote: >> Perhaps we could make walsenders even more like regular backends than what I >> was proposing, so that the replication commands are parsed and executed just >> like regular utility commands. However, that'd require some transaction >> support in walsender, for starters, which seems messy. It might become >> sensible in the future if the replication command set gets even more >> complicated, but it doesn't seem like a good idea at the moment. > It's come up a few times now that people want to run a few queries > either before or after running a base backup. ... > Andres suggested to me the other day we make walsender more like > regular backends. At the time I wasn't sure I agreed, but reading this > it looks like a sensible way to go. That was what I was thinking too, but on reflection there's at least one huge problem: how could we run queries without being connected to a specific database? Which walsender isn't. regards, tom lane
Re: Sharing more infrastructure between walsenders and regular backends (was Re: Switching timeline over streaming replication)
From
Andres Freund
Date:
On Thursday, October 04, 2012 10:58:53 PM Tom Lane wrote: > Simon Riggs <simon@2ndQuadrant.com> writes: > > On 4 October 2012 17:23, Heikki Linnakangas <hlinnakangas@vmware.com> wrote: > >> Perhaps we could make walsenders even more like regular backends than > >> what I was proposing, so that the replication commands are parsed and > >> executed just like regular utility commands. However, that'd require > >> some transaction support in walsender, for starters, which seems messy. > >> It might become sensible in the future if the replication command set > >> gets even more complicated, but it doesn't seem like a good idea at the > >> moment. > > > > It's come up a few times now that people want to run a few queries > > either before or after running a base backup. ... > > Andres suggested to me the other day we make walsender more like > > regular backends. At the time I wasn't sure I agreed, but reading this > > it looks like a sensible way to go. I only went that way after youve disliked my other suggestions ;) > That was what I was thinking too, but on reflection there's at least one > huge problem: how could we run queries without being connected to a > specific database? Which walsender isn't. I had quite some problems with that too. For now I've hacked walsender to connect to the database specified in the connection, not sure whether thats the way to go. Seems to work so far. I wanted to start a thread about this anyway, but as it came up here... The reason "we" (as in logical rep) need a database in the current approach is that we need to access the catalog (in a timetraveling fashion) to know how to decode the data in the wal... The patch I sent two weeks ago does the decoding from inside a normal backend but that was just because I couldn't make walsender work inside a database in time... Greetings, Andres -- Andres Freund http://www.2ndQuadrant.com/ PostgreSQL Development, 24x7 Support, Training & Services
Re: Sharing more infrastructure between walsenders and regular backends (was Re: Switching timeline over streaming replication)
From
Amit Kapila
Date:
On Thursday, October 04, 2012 8:40 PM Heikki Linnakangas wrote: > On 03.10.2012 18:15, Amit Kapila wrote: > > 35.WalSenderMain(void) > > { > > .. > > + if (walsender_shutdown_requested) > > + ereport(FATAL, > > + > (errcode(ERRCODE_ADMIN_SHUTDOWN), > > + errmsg("terminating > > + replication > > connection due to administrator command"))); > > + > > + /* Tell the client that we are ready to receive > > + commands */ > > > > + ReadyForQuery(DestRemote); > > + > > .. > > > > + if (walsender_shutdown_requested) > > + ereport(FATAL, > > + > (errcode(ERRCODE_ADMIN_SHUTDOWN), > > + errmsg("terminating > > + replication > > connection due to administrator command"))); > > + > > > > is it necessary to check walsender_shutdown_requested 2 times in a > > loop, if yes, then can we write comment why it is important to check > > it again. > > The idea was to check for shutdown request before and after the > pq_getbyte() call, because that can block for a long time. > > Looking closer, we don't currently (ie. without this patch) make any > effort to react to SIGTERM in a walsender, while it's waiting for a > command from the client. After starting replication, it does check > walsender_shutdown_requested in the loop, and it's also checked during a > base backup (although only when switching to send next file, which seems > too seldom). This issue is orthogonal to handling timeline changes over > streaming replication, although that patch will make it more important > to handle SIGTERM quickly while waiting for a command, because you stay > in that mode for longer and more often. > > I think walsender needs to share more infrastructure with regular > backends to handle this better. When we first implemented streaming > replication in 9.0, it made sense to implement just the bare minimum > needed to accept the handshake commands before entering the Copy state, > but now that the replication command set has grown to cover base > backups, and fetching timelines with the patch being discussed, we > should bite the bullet and make the command loop more feature-complete > and robust. Certainly there are benefits of making walsender and backend infrastructure similar as mentioned by Simon, Andres and Tom. However shall we not do this as a separate feature along with some other requirements and let switchtimeline patch get committed first as this is altogether a different feature. With Regards, Amit Kapila.
On Thursday, October 04, 2012 7:22 PM Heikki Linnakangas wrote: > > On Wednesday, October 03, 2012 8:45 PM Heikki Linnakangas wrote: > > On Tuesday, October 02, 2012 4:21 PM Heikki Linnakangas wrote: > > > Thanks for the thorough review! I committed the xlog.c refactoring > > patch > > > now. Attached is a new version of the main patch, comments on > specific > > > points below. I didn't adjust the docs per your comments yet, will > do > > > that next. > > > > I have some doubts regarding the comments fixed by you and some more > new > > review comments. > > After this I shall focus majorly towards testing of this Patch. > > > > Testing > ----------- One more test seems to be failed. Apart from this, other tests are passed. 2. a. Master M-1 b. Standby S-1 follows M-1 c. insert 10 records on M-1. verify all records are visible on M-1,S-1 d. Stop S-1 e. insert 2 records on M-1. f. Stop M-1 g. Start S-1 h. Promote S-1 i. Make M-1 recovery.conf suchthat it should connect to S-1 j. Start M-1. Below error comes on M-1 which is expected as M-1 has more data. LOG: database system was shut down at 2012-10-05 16:45:39 IST LOG: entering standby mode LOG: consistentrecovery state reached at 0/176A070 LOG: record with zero length at 0/176A070 LOG: database systemis ready to accept read only connections LOG: streaming replication successfully connected to primary LOG: fetching timeline history file for timeline 2 from primary server LOG: replication terminated by primary server DETAIL: End of WAL reached on timeline 1 LOG: walreceiverended streaming and awaits new instructions LOG: new timeline 2 forked off current database system timeline1 before current recovery point 0/176A070 LOG: re-handshaking at position 0/1000000 on tli 1 LOG: replication terminatedby primary server DETAIL: End of WAL reached on timeline 1 LOG: walreceiver ended streaming and awaitsnew instructions LOG: new timeline 2 forked off current database system timeline 1 before current recovery point 0/176A070 k. Stop M-1. Start M-1. It is able to successfully connect to S-1 which is a problem. l. check in S-1. Records inserted in step-e are not present. m. Now insert records in S-1. M-1 doesn'trecieve any records. On M-1 server following log is getting printed. LOG: out-of-sequence timeline ID 1 (after 2) in log segment 000000020000000000000001, offset 0 LOG: out-of-sequence timeline ID 1 (after 2) in log segment 000000020000000000000001, offset 0 LOG: out-of-sequence timeline ID 1 (after 2) in log segment 000000020000000000000001, offset 0 LOG: out-of-sequence timeline ID 1 (after 2) in log segment 000000020000000000000001, offset 0 LOG: out-of-sequence timeline ID 1 (after 2) in log segment 000000020000000000000001, offset 0With Regards,Amit Kapila.
Re: Promoting a standby during base backup (was Re: Switching timeline over streaming replication)
From
Heikki Linnakangas
Date:
On 04.10.2012 20:07, Fujii Masao wrote: > On Thu, Oct 4, 2012 at 4:59 PM, Heikki Linnakangas >> But I wonder why promoting a standby renders the backup invalid in the first >> place? Fujii, Simon, can you explain that? > > Simon had the same question and I answered it before. > > http://archives.postgresql.org/message-id/CAHGQGwFU04oO8YL5SUcdjVq3BRNi7WtfzTy9wA2kXtZNHicTeA@mail.gmail.com > --------------------------------------- >> You say >> "If the standby is promoted to the master during online backup, the >> backup fails." >> but no explanation of why? >> >> I could work those things out, but I don't want to have to, plus we >> may disagree if I did. > > If the backup succeeds in that case, when we start an archive recovery from that > backup, the recovery needs to cross between two timelines. Which means that > we need to set recovery_target_timeline before starting recovery. Whether > recovery_target_timeline needs to be set or not depends on whether the standby > was promoted during taking the backup. Leaving such a decision to a user seems > fragile. pg_control is backed up last, it would contain the new timeline. No need to set recovery_target_timeline. > pg_basebackup -x ensures that all required files are included in the backup and > we can start recovery without restoring any file from the archive. But > if the standby is promoted during the backup, the timeline history > file would become > an essential file for recovery, but it's not included in the backup. That is true. We could teach it to include the timeline history file, though. > The situation may change if your switching-timeline patch has been committed. > It's useful if we can continue the backup even if the standby is promoted. It wouldn't help with pg_basebackup -x, although it would allow streaming replication to fetch the timeline history file. I guess it's best to keep that restriction for now. But I'll add a TODO item for this. - Heikki
Re: Promoting a standby during base backup (was Re: Switching timeline over streaming replication)
From
Simon Riggs
Date:
On 4 October 2012 18:07, Fujii Masao <masao.fujii@gmail.com> wrote: > On Thu, Oct 4, 2012 at 4:59 PM, Heikki Linnakangas > <hlinnakangas@vmware.com> wrote: >> On 03.10.2012 18:15, Amit Kapila wrote: >>> >>> On Tuesday, October 02, 2012 4:21 PM Heikki Linnakangas wrote: >>>> >>>> Hmm, should a base backup be aborted when the standby is promoted? Does >>>> the promotion render the backup corrupt? >>> >>> >>> I think currently it does so. Pls refer >>> 1. >>> do_pg_stop_backup(char *labelfile, bool waitforarchive) >>> { >>> .. >>> if (strcmp(backupfrom, "standby") == 0&& !backup_started_in_recovery) >>> ereport(ERROR, >>> >>> (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), >>> errmsg("the standby was promoted during >>> online backup"), >>> errhint("This means that the backup >>> being >>> taken is corrupt " >>> "and should not be used. >>> " >>> "Try taking another >>> online >>> backup."))); >>> .. >>> >>> } >> >> >> Okay. I think that check in do_pg_stop_backup() actually already ensures >> that you don't end up with a corrupt backup, even if the standby is promoted >> while a backup is being taken. Admittedly it would be nicer to abort it >> immediately rather than error out at the end. >> >> But I wonder why promoting a standby renders the backup invalid in the first >> place? Fujii, Simon, can you explain that? > > Simon had the same question and I answered it before. > > http://archives.postgresql.org/message-id/CAHGQGwFU04oO8YL5SUcdjVq3BRNi7WtfzTy9wA2kXtZNHicTeA@mail.gmail.com > --------------------------------------- >> You say >> "If the standby is promoted to the master during online backup, the >> backup fails." >> but no explanation of why? >> >> I could work those things out, but I don't want to have to, plus we >> may disagree if I did. > > If the backup succeeds in that case, when we start an archive recovery from that > backup, the recovery needs to cross between two timelines. Which means that > we need to set recovery_target_timeline before starting recovery. Whether > recovery_target_timeline needs to be set or not depends on whether the standby > was promoted during taking the backup. Leaving such a decision to a user seems > fragile. I accepted your answer before, but I think it should be challenged now. This is definitely a time when you really want that backup, so invalidating it for such a weak reason is not useful, even if I understand your original thought. Something that has concerned me is that we don't have an explicit "timeline change record". We *say* we do that at shutdown checkpoints, but that is recorded in the new timeline. So we have the strange situation of changing timeline at two separate places. When we change timeline we really should generate one last WAL on the old timeline that marks an explicit change of timeline and a single exact moment when the timeline change takes place. With PITR we are unable to do that, because any timeline can fork at any point. With smooth switchover we have a special case that is not "anything goes" and there is a good case for not incrementing the timeline at all. This is still a half-formed thought, but at least you should know I'm in the debate. -- Simon Riggs http://www.2ndQuadrant.com/PostgreSQL Development, 24x7 Support, Training & Services
On 06.10.2012 15:58, Amit Kapila wrote: > One more test seems to be failed. Apart from this, other tests are passed. > > 2. a. Master M-1 > b. Standby S-1 follows M-1 > c. insert 10 records on M-1. verify all records are visible on M-1,S-1 > d. Stop S-1 > e. insert 2 records on M-1. > f. Stop M-1 > g. Start S-1 > h. Promote S-1 > i. Make M-1 recovery.conf such that it should connect to S-1 > j. Start M-1. Below error comes on M-1 which is expected as M-1 has more > data. > LOG: database system was shut down at 2012-10-05 16:45:39 IST > LOG: entering standby mode > LOG: consistent recovery state reached at 0/176A070 > LOG: record with zero length at 0/176A070 > LOG: database system is ready to accept read only connections > LOG: streaming replication successfully connected to primary > LOG: fetching timeline history file for timeline 2 from primary > server > LOG: replication terminated by primary server > DETAIL: End of WAL reached on timeline 1 > LOG: walreceiver ended streaming and awaits new instructions > LOG: new timeline 2 forked off current database system timeline 1 > before current recovery point 0/176A070 > LOG: re-handshaking at position 0/1000000 on tli 1 > LOG: replication terminated by primary server > DETAIL: End of WAL reached on timeline 1 > LOG: walreceiver ended streaming and awaits new instructions > LOG: new timeline 2 forked off current database system timeline 1 > before current recovery point 0/176A070 > k. Stop M-1. Start M-1. It is able to successfully connect to S-1 which > is a problem. > l. check in S-1. Records inserted in step-e are not present. > m. Now insert records in S-1. M-1 doesn't recieve any records. On M-1 > server following log is getting printed. > LOG: out-of-sequence timeline ID 1 (after 2) in log segment > 000000020000000000000001, offset 0 > LOG: out-of-sequence timeline ID 1 (after 2) in log segment > 000000020000000000000001, offset 0 > LOG: out-of-sequence timeline ID 1 (after 2) in log segment > 000000020000000000000001, offset 0 > LOG: out-of-sequence timeline ID 1 (after 2) in log segment > 000000020000000000000001, offset 0 > LOG: out-of-sequence timeline ID 1 (after 2) in log segment > 000000020000000000000001, offset 0 Hmm, seems we need to keep track of which timeline we've used to recover before. Before restart, the master correctly notices that timeline 2 forked off earlier in its history, so it cannot recover to that timeline. But after restart the master begins recovery from the previous checkpoint, and because timeline 2 forked off timeline 1 after the checkpoint, it concludes that it can follow that timeline. It doesn't realize that it had some already recovered/flushed some WAL in timeline 1 after the fork-point. Attached is a new version of the patch. I committed the refactoring of XLogPageRead() already, as that was a readability improvement even without this patch. All the reported issues should be fixed now, although I will continue testing this tomorrow. I added various checks that that the correct timeline is followed during recovery. minRecoveryPoint is now accompanied by a timeline ID, so that when we restart recovery, we check that we recover back to minRecoveryPoint along the same timeline as last time. Also, it now checks at beginning of recovery that the checkpoint record comes from the correct timeline. That fixes the problem that you reported above. I also adjusted the error messages on timeline history problems to be more clear. - Heikki
Attachment
On Tuesday, October 09, 2012 10:32 PM Heikki Linnakangas wrote: > On 06.10.2012 15:58, Amit Kapila wrote: > > One more test seems to be failed. Apart from this, other tests are > passed. > > It seems there is one more defect, please check the same Defect: 1. start primary A 2. start standby B following A 3. start cascade standby C following B. 4.Promote standby B. 5. After successful time line switch in cascade standby C, stop C. 6. Restart C, startupis failing with the following error. FATAL: requested timeline 2 does not contain minimum recovery point 0/3000000 on timeline 1 Review: The following statement is present in the hig-availability.sgml file, which is also needs to be modified in the patch. Promoting a cascading standby terminates the immediate downstream replication connections which it serves. This is because the timeline becomes different between standbys, and they can no longer continue replication. The affected standby(s) may reconnect to reestablish streaming replication. I felt some of minor comments are still not handled: 35. +SendTimeLineHistory(TimeLineHistoryCmd *cmd) { .. + fd = PathNameOpenFile(path, O_RDONLY | PG_BINARY, 0666); errorhandling for fd < 0 is missing. 36.+SendTimeLineHistory(TimeLineHistoryCmd *cmd) { .. if (nread <= 0) + ereport(ERROR, + (errcode_for_file_access(), + errmsg("could not read file \"%s\": %m", + path))); FileClose should be done in error case as well. With Regards, Amit Kapila.
On 10 October 2012 15:26, Amit Kapila <amit.kapila@huawei.com> wrote: > On Tuesday, October 09, 2012 10:32 PM Heikki Linnakangas wrote: >> On 06.10.2012 15:58, Amit Kapila wrote: >> > One more test seems to be failed. Apart from this, other tests are >> passed. >> > > It seems there is one more defect, please check the same > Defect: > > 1. start primary A > 2. start standby B following A > 3. start cascade standby C following B. > 4. Promote standby B. > 5. After successful time line switch in cascade standby C, stop C. > 6. Restart C, startup is failing with the following error. > FATAL: requested timeline 2 does not contain minimum recovery point > 0/3000000 on timeline 1 Hmm... I get something different. When I promote standby B, standby C's log shows: LOG: walreceiver ended streaming and awaits new instructions LOG: re-handshaking at position 0/4000000 on tli 1 LOG: fetching timeline history file for timeline 2 from primary server LOG: walreceiver ended streaming and awaits new instructions LOG: new target timeline is 2 Then when I stop then start standby C I get: FATAL: timeline history was not contiguous LOG: startup process (PID 22986) exited with exit code 1 LOG: aborting startup due to startup process failure -- Thom
> -----Original Message----- > From: pgsql-hackers-owner@postgresql.org [mailto:pgsql-hackers- > owner@postgresql.org] On Behalf Of Amit Kapila > Sent: Wednesday, October 10, 2012 7:57 PM > To: 'Heikki Linnakangas' > Cc: 'PostgreSQL-development' > Subject: Re: [HACKERS] Switching timeline over streaming replication > > On Tuesday, October 09, 2012 10:32 PM Heikki Linnakangas wrote: > > On 06.10.2012 15:58, Amit Kapila wrote: > > > One more test seems to be failed. Apart from this, other tests are > > passed. > > > > It seems there is one more defect, please check the same > Defect: > The test is finished from myside. one more issue: Steps to reproduce the defect: 1. Do initdb 2. Set port=2303, wal_level=hot_standby, hot_standby=off, max_walsenders=3 in the postgresql.conf file 3. Enable the replication connection in pg_hba.conf 4. Start the server. Executing the following commands is leading failure. ./pg_basebackup -P -D ../../data_sub -X fetch -p 2303 pg_basebackup: COPY stream ended before last file was finished rm -fr ../../data_sub ./pg_basebackup -P -D ../../data_sub -X fetch -p 2303 pg_basebackup: COPY stream ended before last file was finished The following logs are observed in the server console. ERROR: requested WAL segment 000000000000000000000002 has already been removed ERROR: requested WAL segment 000000000000000000000003 has already been removed With Regards, Amit Kapila.
Heikki Linnakangas wrote: > Attached is a new version of the patch. I committed the refactoring > of XLogPageRead() already, as that was a readability improvement > even without this patch. All the reported issues should be fixed > now, although I will continue testing this tomorrow. I added various > checks that that the correct timeline is followed during recovery. > minRecoveryPoint is now accompanied by a timeline ID, so that when > we restart recovery, we check that we recover back to > minRecoveryPoint along the same timeline as last time. Also, it now > checks at beginning of recovery that the checkpoint record comes > from the correct timeline. That fixes the problem that you reported > above. I also adjusted the error messages on timeline history > problems to be more clear. Heikki, I see Amit found a problem with this patch. I assume you're going to work a bit more on it and submit/commit another version. I'm marking this one Returned with Feedback. Thanks. -- Álvaro Herrera http://www.2ndQuadrant.com/ PostgreSQL Development, 24x7 Support, Training & Services
Here's an updated version of this patch, rebased with master, including the recent replication timeout changes, and some other cleanup. On 12.10.2012 09:34, Amit Kapila wrote: > The test is finished from myside. > > one more issue: > ... > ./pg_basebackup -P -D ../../data_sub -X fetch -p 2303 > pg_basebackup: COPY stream ended before last file was finished Fixed this. However, the test scenario you point to here: http://archives.postgresql.org/message-id/00a801cda6f3$4aba27b0$e02e7710$@kapila@huawei.com still seems to be broken, although I get a different error message now. I'll dig into this.. - Heikki
Attachment
On 10.10.2012 17:26, Amit Kapila wrote: > 36.+SendTimeLineHistory(TimeLineHistoryCmd *cmd) { .. > if (nread<= 0) > + ereport(ERROR, > + (errcode_for_file_access(), > + errmsg("could not read file > \"%s\": %m", > + path))); > > FileClose should be done in error case as well. Hmm, I think you're right. The straightforward fix to just call FileClose() before the ereport()s in that function would not be enough, though. You might run out of memory in pq_sendbytes(), for example, which would throw an error. We could use PG_TRY/CATCH for this, but seems like overkill. Perhaps the simplest fix is to use a global (static) variable for the fd, and clean it up in WalSndErrorCleanup(). This is a fairly general issue, actually. Looking around, I can see at least two similar cases in existing code, with BasicOpenFile, where we will leak file descriptors on error: copy_file: there are several error cases, including out-of-disk space, with no attempt to close the fds. XLogFileInit: again, no attempt to close the file descriptor on failure. This is called at checkpoint from the checkpointer process, to preallocate new xlog files. Given that we haven't heard any complaints of anyone running into these, these are not a big deal in practice, but in theory at least the XLogFileInit leak could lead to serious problems, as it could cause the checkpointer to run out of file descriptors. - Heikki
On 15.11.2012 12:44, Heikki Linnakangas wrote: > Here's an updated version of this patch, rebased with master, including > the recent replication timeout changes, and some other cleanup. > > On 12.10.2012 09:34, Amit Kapila wrote: >> The test is finished from myside. >> >> one more issue: > > ... >> ./pg_basebackup -P -D ../../data_sub -X fetch -p 2303 >> pg_basebackup: COPY stream ended before last file was finished > > Fixed this. > > However, the test scenario you point to here: > http://archives.postgresql.org/message-id/00a801cda6f3$4aba27b0$e02e7710$@kapila@huawei.com > still seems to be broken, although I get a different error message now. > I'll dig into this.. Ok, here's an updated patch again, with that bug fixed. - Heikki
Attachment
On Thursday, November 15, 2012 6:05 PM Heikki Linnakangas wrote: > On 15.11.2012 12:44, Heikki Linnakangas wrote: > > Here's an updated version of this patch, rebased with master, > > including the recent replication timeout changes, and some other > cleanup. > > > > On 12.10.2012 09:34, Amit Kapila wrote: > >> The test is finished from myside. > >> > >> one more issue: > > > ... > >> ./pg_basebackup -P -D ../../data_sub -X fetch -p 2303 > >> pg_basebackup: COPY stream ended before last file was finished > > > > Fixed this. > > > > However, the test scenario you point to here: > > http://archives.postgresql.org/message-id/00a801cda6f3$4aba27b0$e02e77 > > 10$@kapila@huawei.com still seems to be broken, although I get a > > different error message now. > > I'll dig into this.. > > Ok, here's an updated patch again, with that bug fixed. I shall review and test the updated Patch in Commit Fest. With Regards, Amit Kapila.
Heikki Linnakangas <hlinnakangas@vmware.com> writes: > This is a fairly general issue, actually. Looking around, I can see at > least two similar cases in existing code, with BasicOpenFile, where we > will leak file descriptors on error: Um, don't we automatically clean those up during transaction abort? If we don't, we ought to think about that, not about cluttering calling code with certain-to-be-inadequate cleanup in error cases. regards, tom lane
On 15.11.2012 16:55, Tom Lane wrote: > Heikki Linnakangas<hlinnakangas@vmware.com> writes: >> This is a fairly general issue, actually. Looking around, I can see at >> least two similar cases in existing code, with BasicOpenFile, where we >> will leak file descriptors on error: > > Um, don't we automatically clean those up during transaction abort? Not the ones allocated with PathNameOpenFile or BasicOpenFile. Files allocated with AllocateFile() and OpenTemporaryFile() are cleaned up at abort. > If we don't, we ought to think about that, not about cluttering calling > code with certain-to-be-inadequate cleanup in error cases. Agreed. Cleaning up at end-of-xact won't help walsender or other non-backend processes, though, because they don't do transactions. But a top-level ResourceOwner that's reset in the sigsetjmp() cleanup routine would work. - Heikki
On Thursday, November 15, 2012 6:05 PM Heikki Linnakangas wrote: > On 15.11.2012 12:44, Heikki Linnakangas wrote: > > Here's an updated version of this patch, rebased with master, > > including the recent replication timeout changes, and some other > cleanup. > > > > On 12.10.2012 09:34, Amit Kapila wrote: > >> The test is finished from myside. > >> > >> one more issue: > > > ... > >> ./pg_basebackup -P -D ../../data_sub -X fetch -p 2303 > >> pg_basebackup: COPY stream ended before last file was finished > > > > Fixed this. > > > > However, the test scenario you point to here: > > http://archives.postgresql.org/message-id/00a801cda6f3$4aba27b0$e02e77 > > 10$@kapila@huawei.com still seems to be broken, although I get a > > different error message now. > > I'll dig into this.. > > Ok, here's an updated patch again, with that bug fixed. First, I started with test of this Patch. Basic stuff: ------------ - Patch applies OK - Compiles cleanly with no warnings - Regression tests pass except the "standbycheck". From a glance view of the "standbycheck" regression failures are because of sql scripts and expected outputs are little old. The following problems are observed while testing of the patch. Defect-1: 1. start primary A 2. start standby B following A 3. start cascade standby C following B. 4. Promotestandby B. 5. After successful time line switch in cascade standby C, stop C. 6. Restart C, startup is failingwith the following error. LOG: database system was shut down in recovery at 2012-11-16 16:26:29 IST FATAL: requested timeline 2 does not contain minimum recovery point 0/30143A0 on timeline 1 LOG: startup process (PID 415) exited with exit code 1 LOG: aborting startup dueto startup process failure The above defect is already discussed in the following link. http://archives.postgresql.org/message-id/00a801cda6f3$4aba27b0$e02e7710$@ka pila@huawei.com Defect-2: 1. start primary A 2. start standby B following A 3. start cascade standby C following B with 'recovery_target_timeline' option in recovery.conf is disabled. 4. Promote standby B. 5. Cascade Standby C is not able to followthe new master B because of timeline difference. 6. Try to stop the cascade standby C (which is failing and the server is not stopping, observations are as WAL Receiver process is still running and clients are not allowing to connect). The defect-2 is happened only once in my test environment, I will try to reproduce it. With Regards, Amit Kapila.
On 10.10.2012 17:54, Thom Brown wrote: > Hmm... I get something different. When I promote standby B, standby > C's log shows: > > LOG: walreceiver ended streaming and awaits new instructions > LOG: re-handshaking at position 0/4000000 on tli 1 > LOG: fetching timeline history file for timeline 2 from primary server > LOG: walreceiver ended streaming and awaits new instructions > LOG: new target timeline is 2 > > Then when I stop then start standby C I get: > > FATAL: timeline history was not contiguous > LOG: startup process (PID 22986) exited with exit code 1 > LOG: aborting startup due to startup process failure Found & fixed this one. A paren was misplaced in tliOfPointInHistory() function.. On 16.11.2012 16:01, Amit Kapila wrote: > The following problems are observed while testing of the patch. > Defect-1: > > 1. start primary A > 2. start standby B following A > 3. start cascade standby C following B. > 4. Promote standby B. > 5. After successful time line switch in cascade standby C, stop C. > 6. Restart C, startup is failing with the following error. > > LOG: database system was shut down in recovery at 2012-11-16 > 16:26:29 IST > FATAL: requested timeline 2 does not contain minimum recovery point > 0/30143A0 on timeline 1 > LOG: startup process (PID 415) exited with exit code 1 > LOG: aborting startup due to startup process failure > > The above defect is already discussed in the following link. > http://archives.postgresql.org/message-id/00a801cda6f3$4aba27b0$e02e7710$@ka > pila@huawei.com Fixed now, sorry for neglecting this earlier. The problem was that if the primary switched to a new timeline at position X, and the standby followed that switch, on restart it would set minRecoveryPoint to X, and the new > Defect-2: > > 1. start primary A > 2. start standby B following A > 3. start cascade standby C following B with 'recovery_target_timeline' > option in > recovery.conf is disabled. > 4. Promote standby B. > 5. Cascade Standby C is not able to follow the new master B because of > timeline difference. > 6. Try to stop the cascade standby C (which is failing and the > server is not stopping, > observations are as WAL Receiver process is still running and > clients are not allowing to connect). > > The defect-2 is happened only once in my test environment, I will try to > reproduce it. Found it. When restarting the streaming, I reused the WALRCV_STARTING state. But if you then exited recovery, WalRcvRunning() would think that the walreceiver is stuck starting up, because it's been longer than 10 seconds since it was launched and it's still in WALRCV_STARTING state, so it put it into WALRCV_STOPPED state. And walreceiver didn't expect to be put into STOPPED state after having started up successfully already. I added a new explicit WALRCV_RESTARTING state to handle that. In addition to the above bug fixes, there's some small changes since last patch version: * I changed the LOG messages printed in various stages a bit, hopefully making it easier to follow what's happening. Feedback is welcome on when and how we should log, and whether some error messages need clarification. * 'ps' display is updated when the walreceiver enters and exits idle mode * Updated pg_controldata and pg_resetxlog to handle the new minRecoveryPointTLI field I added to the control file. * startup process wakes up walsenders at the end of recovery, so that cascading standbys are notified immediately when the timeline changes. That removes some of the delay in the process. - Heikki
Attachment
On Monday, November 19, 2012 10:54 PM Heikki Linnakangas wrote: > On 10.10.2012 17:54, Thom Brown wrote: > > Hmm... I get something different. When I promote standby B, standby > > C's log shows: > > > > The following problems are observed while testing of the patch. > > Defect-1: > > > > 1. start primary A > > 2. start standby B following A > > 3. start cascade standby C following B. > > 4. Promote standby B. > > 5. After successful time line switch in cascade standby C, stop > C. > > 6. Restart C, startup is failing with the following error. > > > > LOG: database system was shut down in recovery at 2012-11-16 > > 16:26:29 IST > > FATAL: requested timeline 2 does not contain minimum > > recovery point 0/30143A0 on timeline 1 > > LOG: startup process (PID 415) exited with exit code 1 > > LOG: aborting startup due to startup process failure > > > > The above defect is already discussed in the following link. > > http://archives.postgresql.org/message-id/00a801cda6f3$4aba27b0$e02e77 > > 10$@ka > > pila@huawei.com > > Fixed now, sorry for neglecting this earlier. The problem was that if > the primary switched to a new timeline at position X, and the standby > followed that switch, on restart it would set minRecoveryPoint to X, and > the new Not sure, if above is fixed as I don't see any code change for this and in test also it again fails. Below is result of further testing: Some strange logs are observed during testing. Note: Stop the node means doing a smart shutdown. Scenario-1: 1. start primary A 2. start standby B following A 3. start cascade standby C following B. 4. Executethe following commands in the primary A. create table tbl(f int); insert into tbl values(generate_series(1,1000)); 5. Promote standby B. 6. Execute the following commands in the primary B. insert into tbl values(generate_series(1001,2000)); insert into tbl values(generate_series(2001,3000)); The following logs are presents on the following standby C. please check these are proper or not? LOG: restarted WAL streaming at position 0/B000000 on tli 2 LOG: record with zero length at 0/B024C68 LOG: record with zero length at 0/B035528 LOG: out-of-sequence timeline ID 1 (after 2) in log segment 00000002000000000000000B, offset 0 Following two defects are found while testing the new patch. Defect - 1: 1. start primary A 2. start standby B following A 3. start cascade standby C following B. 4. start another standbyD following C. 5. Promote standby B. 6. After successful time line switch in cascade standby C & D, stop D. 7. Restart D, Startup is successful and connecting to standby C. 8. Stop C. 9. Restart C, startup is failing. Defect-2: 1. start primary A 2. start standby B following A 3. start cascade standby C following B. 4. Startanother standby D following C. 5. Execute the following commands in the primary A. create table tbl(f int); insert into tbl values(generate_series(1,1000)); 6. Promote standby B. 7. Execute the following commandsin the primary B. insert into tbl values(generate_series(1001,2000)); insert into tbl values(generate_series(2001,3000)); The following logs are observed on standby C: LOG: restarted WAL streaming at position 0/7000000 on tli 2 ERROR: requested WAL segment 000000020000000000000007has already been removed LOG: record with zero length at 0/7028190 LOG: record with zero length at 0/7048540 LOG: out-of-sequencetimeline ID 1 (after 2) in log segment 000000020000000000000007, offset 0 The following logs are observed on standby D: LOG: restarted WAL streaming at position 0/7000000 on tli 2 LOG: replication terminated by primary server DETAIL: End of WAL reached on timeline 2 FATAL: error reading result of streaming command: ERROR: requested WAL segment 000000020000000000000007 has already been removed LOG: streaming replication successfully connected toprimary 8. Stop standby D normally and restart D. Restart is failing. Code Review ------------------ 1. > Agreed. Cleaning up at end-of-xact won't help walsender or other non-backend processes, though, because they don't do > transactions. But a top-level ResourceOwner that's reset in the sigsetjmp() cleanup routine would work. Do you think cleanup of files be done as part of this patch or should it be handled separately, as it already exists in other paths of code. In that case may be one ToDo item can be added. 2. Also for forbidden_in_wal_sender(firstchar);, instead of handling it as part of each message, isn't it better if we call only once, something like is_command_allowed(firstchar); switch(firstchar) 3. For function WalRcvStreaming(void) { .. if (state == WALRCV_STREAMING || state == WALRCV_STARTING) } I think in above check, it should check the new state as well WALRCV_RESTARTING 4. In function WalReceiverMain(), can we have Log message similar to below for even "started WAL streaming at position", means when it is even first time. if (!first_stream) ereport(LOG, (errmsg("restartedWAL streaming at position %X/%X on tli %u", (uint32) (startpoint >> 32), (uint32) startpoint, startpointTLI))); 5. In function WalReceiverMain(), if (walrcv_startstreaming(startpointTLI, startpoint)) { .. } else ereport(LOG, (errmsg("primary server contains no more WAL on requested timeline %u", startpointTLI))); I think walrcv_startstreaming() can return false even if it cannot start streaming due to non-availablity of WAL at requested timeline. So "no more" in Log message may be misleading in some cases. How about something similar to "primary server doesn't have WAL for requested timeline" 6. *** a/src/bin/pg_controldata/pg_controldata.c --- b/src/bin/pg_controldata/pg_controldata.c *************** *** 237,242 **** main(int argc, char *argv[]) --- 237,244 ---- printf(_("Minimum recovery ending location: %X/%X\n"), (uint32) (ControlFile.minRecoveryPoint>> 32), (uint32) ControlFile.minRecoveryPoint); + printf(_("Recovery ending timeline: %u\n"), + ControlFile.minRecoveryPointTLI); shouldn't Message "Recovery ending timeline" be "Minimum Recovery ending timeline" One Doubt --------------- when pg_receivexlog is receiving xlog file from the standby server, if the standby server got promoted as master the connection between pg_receivexlog and standby server is broken, because of this reason the current xlog file is not renamed as actual file. Whether such a scenario needs any handling? With Regards, Amit Kapila.
On 20.11.2012 15:33, Amit Kapila wrote: > Defect-2: > 1. start primary A > 2. start standby B following A > 3. start cascade standby C following B. > 4. Start another standby D following C. > 5. Execute the following commands in the primary A. > create table tbl(f int); > insert into tbl values(generate_series(1,1000)); > 6. Promote standby B. > 7. Execute the following commands in the primary B. > insert into tbl values(generate_series(1001,2000)); > insert into tbl values(generate_series(2001,3000)); > > The following logs are observed on standby C: > > LOG: restarted WAL streaming at position 0/7000000 on tli 2 > ERROR: requested WAL segment 000000020000000000000007 has already been > removed > LOG: record with zero length at 0/7028190 > LOG: record with zero length at 0/7048540 > LOG: out-of-sequence timeline ID 1 (after 2) in log segment > 000000020000000000000007, offset 0 Hmm, this one is actually a pre-existing bug. There's a sanity check that the sequence of timeline IDs that are seen in the XLOG page headers doesn't go backwards. In other words, if the last XLOG page that was read had timeline id X, the next page must have a tli >= X. The startup process keeps track of the last seen timeline id in lastPageTLI. In standby_mode, when the startup process is reading from a pre-existing file in pg_xlog (typically put there by streaming replication) and it reaches the end of valid WAL (marked by an error in decoding it, ie. "record with zero length" in your case), it sleeps for five seconds and retries. At retry, the WAL file is re-opened, and as part of sanity checking it, the first page header in the file is validated. Now, if there was a timeline change in the current WAL segment, and we've already replayed past that point, lastPageTLI will already be set to the new TLI, but the first page on the file contains the old TLI. When the file is re-opened, and the first page is validated, you get the error. The fix is quite straightforward: we should refrain from checking the TLI when we re-open a WAL file. Or better yet, compare it against the TLI we saw at the beginning of the last WAL segment, not the last WAL page. I propose the attached patch (against 9.2) to fix that. This should be backpatched to 9.0, where standby_mode was introduced. The code was the same in 8.4, too, but AFAICS there was no problem there because 8.4 never tried to re-open the same WAL segment after replaying some of it. - Heikki
Attachment
On Wednesday, November 21, 2012 11:36 PM Heikki Linnakangas wrote: > On 20.11.2012 15:33, Amit Kapila wrote: > > Defect-2: > > 1. start primary A > > 2. start standby B following A > > 3. start cascade standby C following B. > > 4. Start another standby D following C. > > 5. Execute the following commands in the primary A. > > create table tbl(f int); > > insert into tbl values(generate_series(1,1000)); > > 6. Promote standby B. > > 7. Execute the following commands in the primary B. > > insert into tbl values(generate_series(1001,2000)); > > insert into tbl values(generate_series(2001,3000)); > > > > The following logs are observed on standby C: > > > > LOG: restarted WAL streaming at position 0/7000000 on tli 2 > > ERROR: requested WAL segment 000000020000000000000007 has > > already been removed > > LOG: record with zero length at 0/7028190 > > LOG: record with zero length at 0/7048540 > > LOG: out-of-sequence timeline ID 1 (after 2) in log segment > > 000000020000000000000007, offset 0 > > I propose the attached patch (against 9.2) to fix that. This should be > backpatched to 9.0, where standby_mode was introduced. The code was the > same in 8.4, too, but AFAICS there was no problem there because 8.4 > never tried to re-open the same WAL segment after replaying some of it. Fixed. With Regards, Amit Kapila.
Plugging fd leaks (was Re: Switching timeline over streaming replication)
From
Heikki Linnakangas
Date:
On 15.11.2012 17:16, Heikki Linnakangas wrote: > On 15.11.2012 16:55, Tom Lane wrote: >> Heikki Linnakangas<hlinnakangas@vmware.com> writes: >>> This is a fairly general issue, actually. Looking around, I can see at >>> least two similar cases in existing code, with BasicOpenFile, where we >>> will leak file descriptors on error: >> >> Um, don't we automatically clean those up during transaction abort? > > Not the ones allocated with PathNameOpenFile or BasicOpenFile. Files > allocated with AllocateFile() and OpenTemporaryFile() are cleaned up at > abort. > >> If we don't, we ought to think about that, not about cluttering calling >> code with certain-to-be-inadequate cleanup in error cases. > > Agreed. Cleaning up at end-of-xact won't help walsender or other > non-backend processes, though, because they don't do transactions. But a > top-level ResourceOwner that's reset in the sigsetjmp() cleanup routine > would work. This is what I came up with. It adds a new function, OpenFile, that returns a raw file descriptor like BasicOpenFile, but the file descriptor is associated with the current subtransaction and automatically closed at end-of-xact, in the same way that AllocateFile and AllocateDir work. In other words, OpenFile is to open() what AllocateFile is to fopen(). BasicOpenFile is unchanged, it returns a raw fd and it's solely the caller's responsibility to close it, but many of the places that used to call BasicOpenFile now use the safer OpenFile function instead. This patch plugs three existing fd (or virtual fd) leaks: 1. copy_file() - fixed by by using OpenFile instead of BasicOpenFile 2. XLogFileLinit() - fixed by adding close() calls to the error cases. Can't use OpenFile here because the fd is supposed to persist over transaction boundaries. 3. lo_import/lo_export - fixed by using OpenFile instead of PathNameOpenFile. In addition, this replaces many BasicOpenFile() calls with OpenFile() that were not leaking, because the code meticulously closed the file on error. That wasn't strictly necessary, but IMHO it's good for robustness. One thing I'm not too fond of is the naming. I'm calling the new functions OpenFile and CloseFile. There's some danger of confusion there, as the function to close a virtual file opened with PathNameOpenFile is called FileClose. OpenFile is really the same kind of operation as AllocateFile and AllocateDir, but returns an unbuffered fd. So it would be nice if it was called AllocateSomething, too. But AllocateFile is already taken. And I don't much like the Allocate* naming for these anyway, you really would expect the name to contain "open". Do we want to backpatch this? We've had zero complaints, but this seems fairly safe to backpatch, and at least the leak in copy_file() can be quite annoying. If you run out of disk space in CREATE DATABASE, the target file is kept open even though it's deleted, so the space isn't reclaimed until you disconnect. - Heikki
Attachment
Re: Plugging fd leaks (was Re: Switching timeline over streaming replication)
From
Amit Kapila
Date:
On Friday, November 23, 2012 7:03 PM Heikki Linnakangas > On 15.11.2012 17:16, Heikki Linnakangas wrote: > > On 15.11.2012 16:55, Tom Lane wrote: > >> Heikki Linnakangas<hlinnakangas@vmware.com> writes: > >>> This is a fairly general issue, actually. Looking around, I can see > >>> at least two similar cases in existing code, with BasicOpenFile, > >>> where we will leak file descriptors on error: > >> > >> Um, don't we automatically clean those up during transaction abort? > > > > Not the ones allocated with PathNameOpenFile or BasicOpenFile. Files > > allocated with AllocateFile() and OpenTemporaryFile() are cleaned up > > at abort. > > > >> If we don't, we ought to think about that, not about cluttering > >> calling code with certain-to-be-inadequate cleanup in error cases. > > > > Agreed. Cleaning up at end-of-xact won't help walsender or other > > non-backend processes, though, because they don't do transactions. But > > a top-level ResourceOwner that's reset in the sigsetjmp() cleanup > > routine would work. > > This is what I came up with. It adds a new function, OpenFile, that > returns a raw file descriptor like BasicOpenFile, but the file > descriptor is associated with the current subtransaction and > automatically closed at end-of-xact, in the same way that AllocateFile > and AllocateDir work. In other words, OpenFile is to open() what > AllocateFile is to fopen(). BasicOpenFile is unchanged, it returns a raw > fd and it's solely the caller's responsibility to close it, but many of > the places that used to call BasicOpenFile now use the safer OpenFile > function instead. > > This patch plugs three existing fd (or virtual fd) leaks: > > 1. copy_file() - fixed by by using OpenFile instead of BasicOpenFile 2. > XLogFileLinit() - fixed by adding close() calls to the error cases. > Can't use OpenFile here because the fd is supposed to persist over > transaction boundaries. > 3. lo_import/lo_export - fixed by using OpenFile instead of > PathNameOpenFile. I have gone through the patch and find it okay except for one minor suggestion 1. Can we put below log in OpenFile as well + DO_DB(elog(LOG, "CloseFile: Allocated %d", numAllocatedDescs)); > One thing I'm not too fond of is the naming. I'm calling the new > functions OpenFile and CloseFile. There's some danger of confusion > there, as the function to close a virtual file opened with > PathNameOpenFile is called FileClose. OpenFile is really the same kind > of operation as AllocateFile and AllocateDir, but returns an unbuffered > fd. So it would be nice if it was called AllocateSomething, too. But > AllocateFile is already taken. And I don't much like the Allocate* > naming for these anyway, you really would expect the name to contain > "open". OpenFileInTrans OpenTransactionAwareFile In anycase OpenFile is also okay. I have one usecase in feature (SQL Command to edit postgresql.conf) very similar to OpenFile/CloseFile, but I want that when CloseFile is called from abort, it should remove(unlink) the file as well and during open it has to retry few times if open is not success. I have following options: 1. Extend OpenFile/CloseFile or PathNameOpenFile 2. Write new functions similar to OpenFile/CloseFile, something like OpenConfLockFile/CloseConfLockFile 3. Use OpenFile/CloseFile and handle my specific case with PG_TRY .. PG_CATCH Any suggestions? With Regards, Amit Kapila.
Re: Plugging fd leaks (was Re: Switching timeline over streaming replication)
From
Heikki Linnakangas
Date:
On 26.11.2012 14:53, Amit Kapila wrote: > I have one usecase in feature (SQL Command to edit postgresql.conf) very > similar to OpenFile/CloseFile, but I want that when CloseFile is called from > abort, it should remove(unlink) the file as well and during open it has to > retry few times if open is not success. > I have following options: > 1. Extend OpenFile/CloseFile or PathNameOpenFile > 2. Write new functions similar to OpenFile/CloseFile, something like > OpenConfLockFile/CloseConfLockFile > 3. Use OpenFile/CloseFile and handle my specific case with PG_TRY .. > PG_CATCH > > Any suggestions? Hmm, if it's just for locking purposes, how about using a lwlock or a heavy-weight lock instead? - Heikki
Re: Plugging fd leaks (was Re: Switching timeline over streaming replication)
From
Amit Kapila
Date:
On Monday, November 26, 2012 7:01 PM Heikki Linnakangas wrote: > On 26.11.2012 14:53, Amit Kapila wrote: > > I have one usecase in feature (SQL Command to edit postgresql.conf) > very > > similar to OpenFile/CloseFile, but I want that when CloseFile is > called from > > abort, it should remove(unlink) the file as well and during open it > has to > > retry few times if open is not success. > > I have following options: > > 1. Extend OpenFile/CloseFile or PathNameOpenFile > > 2. Write new functions similar to OpenFile/CloseFile, something like > > OpenConfLockFile/CloseConfLockFile > > 3. Use OpenFile/CloseFile and handle my specific case with PG_TRY .. > > PG_CATCH > > > > Any suggestions? > > Hmm, if it's just for locking purposes, how about using a lwlock or a > heavy-weight lock instead? Its not only for lock, the main idea is that we create temp file and write modified configuration in that temp file. In end if it's success, then we rename temp file to .conf file but if it error out then at abort we need to delete temp file. So in short, main point is to close/rename the file in case of success (at end of command) and remove in case of abort. With Regards, Amit Kapila.
Amit Kapila <amit.kapila@huawei.com> writes: > On Monday, November 26, 2012 7:01 PM Heikki Linnakangas wrote: >> Hmm, if it's just for locking purposes, how about using a lwlock or a >> heavy-weight lock instead? > Its not only for lock, the main idea is that we create temp file and write > modified configuration in that temp file. > In end if it's success, then we rename temp file to .conf file but if it > error out then at abort we need to delete temp file. > So in short, main point is to close/rename the file in case of success (at > end of command) and remove in case of abort. I'd go with the TRY/CATCH solution. It would be worth extending the fd.c infrastructure if there were multiple users of the feature, but there are not, nor do I see likely new candidates on the horizon. regards, tom lane
Re: Plugging fd leaks (was Re: Switching timeline over streaming replication)
From
Heikki Linnakangas
Date:
On 26.11.2012 14:53, Amit Kapila wrote: > On Friday, November 23, 2012 7:03 PM Heikki Linnakangas >> This is what I came up with. It adds a new function, OpenFile, that >> returns a raw file descriptor like BasicOpenFile, but the file >> descriptor is associated with the current subtransaction and >> automatically closed at end-of-xact, in the same way that AllocateFile >> and AllocateDir work. In other words, OpenFile is to open() what >> AllocateFile is to fopen(). BasicOpenFile is unchanged, it returns a raw >> fd and it's solely the caller's responsibility to close it, but many of >> the places that used to call BasicOpenFile now use the safer OpenFile >> function instead. >> >> This patch plugs three existing fd (or virtual fd) leaks: >> >> 1. copy_file() - fixed by by using OpenFile instead of BasicOpenFile 2. >> XLogFileLinit() - fixed by adding close() calls to the error cases. >> Can't use OpenFile here because the fd is supposed to persist over >> transaction boundaries. >> 3. lo_import/lo_export - fixed by using OpenFile instead of >> PathNameOpenFile. > > I have gone through the patch and find it okay except for one minor > suggestion > 1. Can we put below log in OpenFile as well > + DO_DB(elog(LOG, "CloseFile: Allocated %d", numAllocatedDescs)); Thanks. Added that and committed. I didn't dare to backpatch this, even though it could be fairly easily backpatched. The leaks exist in older versions too, but since they're extremely rare (zero complaints from the field and it's been like that forever), I didn't want to take the risk. Maybe later, after this has had more testing in master. >> One thing I'm not too fond of is the naming. I'm calling the new >> functions OpenFile and CloseFile. There's some danger of confusion >> there, as the function to close a virtual file opened with >> PathNameOpenFile is called FileClose. OpenFile is really the same kind >> of operation as AllocateFile and AllocateDir, but returns an unbuffered >> fd. So it would be nice if it was called AllocateSomething, too. But >> AllocateFile is already taken. And I don't much like the Allocate* >> naming for these anyway, you really would expect the name to contain >> "open". > > OpenFileInTrans > OpenTransactionAwareFile > > In anycase OpenFile is also okay. I ended up calling the functions OpenTransientFile and CloseTransientFile. Windows has a library function called "OpenFile", so that was a pretty bad choice after all. - Heikki
Is this patch available in version 9.2.1 ? Senthil -- View this message in context: http://postgresql.1045698.n5.nabble.com/Switching-timeline-over-streaming-replication-tp5723547p5734744.html Sent from the PostgreSQL - hackers mailing list archive at Nabble.com.
On 03.12.2012 14:21, senthilnathan wrote: > Is this patch available in version 9.2.1 ? Nope, this is for 9.3. - Heikki
After some diversions to fix bugs and refactor existing code, I've committed a couple of small parts of this patch, which just add some sanity checks to notice incorrect PITR scenarios. Here's a new version of the main patch based on current HEAD. - Heikki
Attachment
On Tuesday, December 04, 2012 10:01 PM Heikki Linnakangas wrote: > After some diversions to fix bugs and refactor existing code, I've > committed a couple of small parts of this patch, which just add some > sanity checks to notice incorrect PITR scenarios. Here's a new version > of the main patch based on current HEAD. After testing with the new patch, the following problems are observed. Defect - 1: 1. start primary A 2. start standby B following A 3. start cascade standby C following B. 4. start another standbyD following C. 5. Promote standby B. 6. After successful time line switch in cascade standby C & D, stop D. 7. Restart D, Startup is successful and connecting to standby C. 8. Stop C. 9. Restart C, startup is failing. Defect-2: 1. start primary A 2. start standby B following A 3. start cascade standby C following B. 4. Startanother standby D following C. 5. Execute the following commands in the primary A. create table tbl(f int); insert into tbl values(generate_series(1,1000)); 6. Promote standby B. 7. Execute the following commandsin the primary B. insert into tbl values(generate_series(1001,2000)); insert into tbl values(generate_series(2001,3000)); 8. Stop standby D normally and restart D. Restart is failing. 9. Stopstandby C normally and restart C. Restart is failing. Note: Stop the node means doing a smart shutdown. With Regards, Amit Kapila.
On 05.12.2012 14:32, Amit Kapila wrote: > On Tuesday, December 04, 2012 10:01 PM Heikki Linnakangas wrote: >> After some diversions to fix bugs and refactor existing code, I've >> committed a couple of small parts of this patch, which just add some >> sanity checks to notice incorrect PITR scenarios. Here's a new version >> of the main patch based on current HEAD. > > After testing with the new patch, the following problems are observed. > > Defect - 1: > > 1. start primary A > 2. start standby B following A > 3. start cascade standby C following B. > 4. start another standby D following C. > 5. Promote standby B. > 6. After successful time line switch in cascade standby C& D, stop D. > 7. Restart D, Startup is successful and connecting to standby C. > 8. Stop C. > 9. Restart C, startup is failing. Ok, the error I get in that scenario is: C 2012-12-05 19:55:43.840 EET 9283 FATAL: requested timeline 2 does not contain minimum recovery point 0/3023F08 on timeline 1 C 2012-12-05 19:55:43.841 EET 9282 LOG: startup process (PID 9283) exited with exit code 1 C 2012-12-05 19:55:43.841 EET 9282 LOG: aborting startup due to startup process failure It seems that the commits I made to master already: http://archives.postgresql.org/pgsql-committers/2012-12/msg00116.php http://archives.postgresql.org/pgsql-committers/2012-12/msg00111.php were a few bricks shy of a load. The problem is that if recovery stops at a checkpoint record that changes timeline, so that minRecoveryPoint points to the end of the checkpoint record, we still record the old TLI as the TLI of minRecoveryPoint. This is because 1) there's a silly bug in the patch; replayEndTLI is not updated along with replayEndRecPtr. But even after fixing that, we're not good. The problem is that replayEndRecPtr is currently updated *before* calling the redo function. replayEndRecPtr is what becomes minRecoveryPoint when XLogFlush() is called. If the record is a checkpoint record, redoing it will switch recovery to the new timeline, but replayEndTLI will not be updated until the next record. IOW, as far as minRecoveryPoint is concerned, a checkpoint record that switches timeline is considered to be part of the old timeline. But when a server is promoted and a new timeline is created, the checkpoint record is considered to be part of the new timeline; that's what we write in the page header and in the control file. That mismatch causes the error. I'd like to fix this by always treating the checkpoint record to be part of the new timeline. That feels more correct. The most straightforward way to implement that would be to peek at the xlog record before updating replayEndRecPtr and replayEndTLI. If it's a checkpoint record that changes TLI, set replayEndTLI to the new timeline before calling the redo-function. But it's a bit of a modularity violation to peek into the record like that. Or we could just revert the sanity check at beginning of recovery that throws the "requested timeline 2 does not contain minimum recovery point 0/3023F08 on timeline 1" error. The error I added to redo of checkpoint record that says "unexpected timeline ID %u in checkpoint record, before reaching minimum recovery point %X/%X on timeline %u" checks basically the same thing, but at a later stage. However, the way minRecoveryPointTLI is updated still seems wrong to me, so I'd like to fix that. I'm thinking of something like the attached (with some more comments before committing). Thoughts? - Heikki
Attachment
On Thursday, December 06, 2012 12:53 AM Heikki Linnakangas wrote: > On 05.12.2012 14:32, Amit Kapila wrote: > > On Tuesday, December 04, 2012 10:01 PM Heikki Linnakangas wrote: > >> After some diversions to fix bugs and refactor existing code, I've > >> committed a couple of small parts of this patch, which just add some > >> sanity checks to notice incorrect PITR scenarios. Here's a new > >> version of the main patch based on current HEAD. > > > > After testing with the new patch, the following problems are observed. > > > > Defect - 1: > > > > 1. start primary A > > 2. start standby B following A > > 3. start cascade standby C following B. > > 4. start another standby D following C. > > 5. Promote standby B. > > 6. After successful time line switch in cascade standby C& D, > stop D. > > 7. Restart D, Startup is successful and connecting to standby C. > > 8. Stop C. > > 9. Restart C, startup is failing. > > Ok, the error I get in that scenario is: > > C 2012-12-05 19:55:43.840 EET 9283 FATAL: requested timeline 2 does not > contain minimum recovery point 0/3023F08 on timeline 1 C 2012-12-05 > 19:55:43.841 EET 9282 LOG: startup process (PID 9283) exited with exit > code 1 C 2012-12-05 19:55:43.841 EET 9282 LOG: aborting startup due to > startup process failure > > > That mismatch causes the error. I'd like to fix this by always treating > the checkpoint record to be part of the new timeline. That feels more > correct. The most straightforward way to implement that would be to peek > at the xlog record before updating replayEndRecPtr and replayEndTLI. If > it's a checkpoint record that changes TLI, set replayEndTLI to the new > timeline before calling the redo-function. But it's a bit of a > modularity violation to peek into the record like that. > > Or we could just revert the sanity check at beginning of recovery that > throws the "requested timeline 2 does not contain minimum recovery point > 0/3023F08 on timeline 1" error. The error I added to redo of checkpoint > record that says "unexpected timeline ID %u in checkpoint record, before > reaching minimum recovery point %X/%X on timeline %u" checks basically > the same thing, but at a later stage. However, the way > minRecoveryPointTLI is updated still seems wrong to me, so I'd like to > fix that. > > I'm thinking of something like the attached (with some more comments > before committing). Thoughts? This has fixed the problem reported. However, I am not able to think will there be any problem if we remove check "requested timeline 2 does not contain minimum recovery point > 0/3023F08 on timeline 1" at beginning of recovery and just update replayEndTLI with ThisTimeLineID? With Regards, Amit Kapila.
On 06.12.2012 15:39, Amit Kapila wrote: > On Thursday, December 06, 2012 12:53 AM Heikki Linnakangas wrote: >> On 05.12.2012 14:32, Amit Kapila wrote: >>> On Tuesday, December 04, 2012 10:01 PM Heikki Linnakangas wrote: >>>> After some diversions to fix bugs and refactor existing code, I've >>>> committed a couple of small parts of this patch, which just add some >>>> sanity checks to notice incorrect PITR scenarios. Here's a new >>>> version of the main patch based on current HEAD. >>> >>> After testing with the new patch, the following problems are observed. >>> >>> Defect - 1: >>> >>> 1. start primary A >>> 2. start standby B following A >>> 3. start cascade standby C following B. >>> 4. start another standby D following C. >>> 5. Promote standby B. >>> 6. After successful time line switch in cascade standby C& D, >> stop D. >>> 7. Restart D, Startup is successful and connecting to standby C. >>> 8. Stop C. >>> 9. Restart C, startup is failing. >> >> Ok, the error I get in that scenario is: >> >> C 2012-12-05 19:55:43.840 EET 9283 FATAL: requested timeline 2 does not >> contain minimum recovery point 0/3023F08 on timeline 1 C 2012-12-05 >> 19:55:43.841 EET 9282 LOG: startup process (PID 9283) exited with exit >> code 1 C 2012-12-05 19:55:43.841 EET 9282 LOG: aborting startup due to >> startup process failure >> > >> >> That mismatch causes the error. I'd like to fix this by always treating >> the checkpoint record to be part of the new timeline. That feels more >> correct. The most straightforward way to implement that would be to peek >> at the xlog record before updating replayEndRecPtr and replayEndTLI. If >> it's a checkpoint record that changes TLI, set replayEndTLI to the new >> timeline before calling the redo-function. But it's a bit of a >> modularity violation to peek into the record like that. >> >> Or we could just revert the sanity check at beginning of recovery that >> throws the "requested timeline 2 does not contain minimum recovery point >> 0/3023F08 on timeline 1" error. The error I added to redo of checkpoint >> record that says "unexpected timeline ID %u in checkpoint record, before >> reaching minimum recovery point %X/%X on timeline %u" checks basically >> the same thing, but at a later stage. However, the way >> minRecoveryPointTLI is updated still seems wrong to me, so I'd like to >> fix that. >> >> I'm thinking of something like the attached (with some more comments >> before committing). Thoughts? > > This has fixed the problem reported. > However, I am not able to think will there be any problem if we remove check > "requested timeline 2 does not contain minimum recovery point >> 0/3023F08 on timeline 1" at beginning of recovery and just update > replayEndTLI with ThisTimeLineID? Well, it seems wrong for the control file to contain a situation like this: pg_control version number: 932 Catalog version number: 201211281 Database system identifier: 5819228770976387006 Database cluster state: shut down in recovery pg_control last modified: pe 7. joulukuuta 2012 17.39.57 Latest checkpoint location: 0/3023EA8 Prior checkpoint location: 0/2000060 Latest checkpoint's REDO location: 0/3023EA8 Latest checkpoint's REDO WAL file: 000000020000000000000003 Latest checkpoint's TimeLineID: 2 ... Time of latest checkpoint: pe 7. joulukuuta 2012 17.39.49 Min recovery ending location: 0/3023F08 Min recovery ending loc's timeline: 1 Note the latest checkpoint location and its TimelineID, and compare them with the min recovery ending location. The min recovery ending location is ahead of latest checkpoint's location; the min recovery ending location actually points to the end of the checkpoint record. But how come the min recovery ending location's timeline is 1, while the checkpoint record's timeline is 2. Now maybe that would happen to work if remove the sanity check, but it still seems horribly confusing. I'm afraid that discrepancy will come back to haunt us later if we leave it like that. So I'd like to fix that. Mulling over this for some more, I propose the attached patch. With the patch, we peek into the checkpoint record, and actually perform the timeline switch (by changing ThisTimeLineID) before replaying it. That way the checkpoint record is really considered to be on the new timeline for all purposes. At the moment, the only difference that makes in practice is that we set replayEndTLI, and thus minRecoveryPointTLI, to the new TLI, but it feels logically more correct to do it that way. - Heikki
Attachment
> From: Heikki Linnakangas [mailto:hlinnakangas@vmware.com] > Sent: Friday, December 07, 2012 9:22 PM > To: Amit Kapila > Cc: 'PostgreSQL-development'; 'Thom Brown' > Subject: Re: [HACKERS] Switching timeline over streaming replication > > On 06.12.2012 15:39, Amit Kapila wrote: > > On Thursday, December 06, 2012 12:53 AM Heikki Linnakangas wrote: > >> On 05.12.2012 14:32, Amit Kapila wrote: > >>> On Tuesday, December 04, 2012 10:01 PM Heikki Linnakangas wrote: > >>>> After some diversions to fix bugs and refactor existing code, I've > >>>> committed a couple of small parts of this patch, which just add > >>>> some sanity checks to notice incorrect PITR scenarios. Here's a new > >>>> version of the main patch based on current HEAD. > >>> > >>> After testing with the new patch, the following problems are > observed. > >>> > >>> Defect - 1: > >>> > >>> 1. start primary A > >>> 2. start standby B following A > >>> 3. start cascade standby C following B. > >>> 4. start another standby D following C. > >>> 5. Promote standby B. > >>> 6. After successful time line switch in cascade standby C& > D, > >> stop D. > >>> 7. Restart D, Startup is successful and connecting to standby > C. > >>> 8. Stop C. > >>> 9. Restart C, startup is failing. > >> > >> Ok, the error I get in that scenario is: > >> > >> C 2012-12-05 19:55:43.840 EET 9283 FATAL: requested timeline 2 does > >> not contain minimum recovery point 0/3023F08 on timeline 1 C > >> 2012-12-05 > >> 19:55:43.841 EET 9282 LOG: startup process (PID 9283) exited with > >> exit code 1 C 2012-12-05 19:55:43.841 EET 9282 LOG: aborting startup > >> due to startup process failure > >> > > > >> > Well, it seems wrong for the control file to contain a situation like > this: > > pg_control version number: 932 > Catalog version number: 201211281 > Database system identifier: 5819228770976387006 > Database cluster state: shut down in recovery > pg_control last modified: pe 7. joulukuuta 2012 17.39.57 > Latest checkpoint location: 0/3023EA8 > Prior checkpoint location: 0/2000060 > Latest checkpoint's REDO location: 0/3023EA8 > Latest checkpoint's REDO WAL file: 000000020000000000000003 > Latest checkpoint's TimeLineID: 2 > ... > Time of latest checkpoint: pe 7. joulukuuta 2012 17.39.49 > Min recovery ending location: 0/3023F08 > Min recovery ending loc's timeline: 1 > > Note the latest checkpoint location and its TimelineID, and compare them > with the min recovery ending location. The min recovery ending location > is ahead of latest checkpoint's location; the min recovery ending > location actually points to the end of the checkpoint record. But how > come the min recovery ending location's timeline is 1, while the > checkpoint record's timeline is 2. > > Now maybe that would happen to work if remove the sanity check, but it > still seems horribly confusing. I'm afraid that discrepancy will come > back to haunt us later if we leave it like that. So I'd like to fix > that. > > Mulling over this for some more, I propose the attached patch. With the > patch, we peek into the checkpoint record, and actually perform the > timeline switch (by changing ThisTimeLineID) before replaying it. That > way the checkpoint record is really considered to be on the new timeline > for all purposes. At the moment, the only difference that makes in > practice is that we set replayEndTLI, and thus minRecoveryPointTLI, to > the new TLI, but it feels logically more correct to do it that way. This has fixed both the problems reported in below link: http://archives.postgresql.org/pgsql-hackers/2012-12/msg00267.php The code is also fine. With Regards, Amit Kapila.
Heikki, Tested this on yesterday's snapshot. Worked great. Test: 4 Ubuntu 10.04 LTS Cloud Servers (GoGrid) Configuration: Compiled 9.3(12-12-12)with: pg_stat_statements, citext, ISN, btree_gist, pl/perl Setup Test: Master-Master Replicated to: master-replica using pg_basebackup -x.No archiving. Master-Replicareplicated to Replica-Replica1 and Replica-Replica2using pg_basebackup -x All came up on first try, with no issues. Ran customized pgbench (with waits); lag time to cascading replicas was < 1 second. Failover Test: 1. started customized pgbench on master-master. 2. shut down master-master (-fast) 3. promoted master-replica to new master 4. restarted custom pgbench, at master-replica Result: Replication to replica-replica1,2 working fine, no interruptions in existing connections to replica-replicas. Now I wanna test a chain of cascading replicas ... how far can we chain these? -- Josh Berkus PostgreSQL Experts Inc. http://pgexperts.com
On Sat, Dec 8, 2012 at 12:51 AM, Heikki Linnakangas <hlinnakangas@vmware.com> wrote: > On 06.12.2012 15:39, Amit Kapila wrote: >> >> On Thursday, December 06, 2012 12:53 AM Heikki Linnakangas wrote: >>> >>> On 05.12.2012 14:32, Amit Kapila wrote: >>>> >>>> On Tuesday, December 04, 2012 10:01 PM Heikki Linnakangas wrote: >>>>> >>>>> After some diversions to fix bugs and refactor existing code, I've >>>>> committed a couple of small parts of this patch, which just add some >>>>> sanity checks to notice incorrect PITR scenarios. Here's a new >>>>> version of the main patch based on current HEAD. >>>> >>>> >>>> After testing with the new patch, the following problems are observed. >>>> >>>> Defect - 1: >>>> >>>> 1. start primary A >>>> 2. start standby B following A >>>> 3. start cascade standby C following B. >>>> 4. start another standby D following C. >>>> 5. Promote standby B. >>>> 6. After successful time line switch in cascade standby C& D, >>> >>> stop D. >>>> >>>> 7. Restart D, Startup is successful and connecting to standby C. >>>> 8. Stop C. >>>> 9. Restart C, startup is failing. >>> >>> >>> Ok, the error I get in that scenario is: >>> >>> C 2012-12-05 19:55:43.840 EET 9283 FATAL: requested timeline 2 does not >>> contain minimum recovery point 0/3023F08 on timeline 1 C 2012-12-05 >>> 19:55:43.841 EET 9282 LOG: startup process (PID 9283) exited with exit >>> code 1 C 2012-12-05 19:55:43.841 EET 9282 LOG: aborting startup due to >>> startup process failure >>> >> >>> >>> That mismatch causes the error. I'd like to fix this by always treating >>> the checkpoint record to be part of the new timeline. That feels more >>> correct. The most straightforward way to implement that would be to peek >>> at the xlog record before updating replayEndRecPtr and replayEndTLI. If >>> it's a checkpoint record that changes TLI, set replayEndTLI to the new >>> timeline before calling the redo-function. But it's a bit of a >>> modularity violation to peek into the record like that. >>> >>> Or we could just revert the sanity check at beginning of recovery that >>> throws the "requested timeline 2 does not contain minimum recovery point >>> 0/3023F08 on timeline 1" error. The error I added to redo of checkpoint >>> record that says "unexpected timeline ID %u in checkpoint record, before >>> reaching minimum recovery point %X/%X on timeline %u" checks basically >>> the same thing, but at a later stage. However, the way >>> minRecoveryPointTLI is updated still seems wrong to me, so I'd like to >>> fix that. >>> >>> I'm thinking of something like the attached (with some more comments >>> before committing). Thoughts? >> >> >> This has fixed the problem reported. >> However, I am not able to think will there be any problem if we remove >> check >> "requested timeline 2 does not contain minimum recovery point >>> >>> 0/3023F08 on timeline 1" at beginning of recovery and just update >> >> replayEndTLI with ThisTimeLineID? > > > Well, it seems wrong for the control file to contain a situation like this: > > pg_control version number: 932 > Catalog version number: 201211281 > Database system identifier: 5819228770976387006 > Database cluster state: shut down in recovery > pg_control last modified: pe 7. joulukuuta 2012 17.39.57 > Latest checkpoint location: 0/3023EA8 > Prior checkpoint location: 0/2000060 > Latest checkpoint's REDO location: 0/3023EA8 > Latest checkpoint's REDO WAL file: 000000020000000000000003 > Latest checkpoint's TimeLineID: 2 > ... > Time of latest checkpoint: pe 7. joulukuuta 2012 17.39.49 > Min recovery ending location: 0/3023F08 > Min recovery ending loc's timeline: 1 > > Note the latest checkpoint location and its TimelineID, and compare them > with the min recovery ending location. The min recovery ending location is > ahead of latest checkpoint's location; the min recovery ending location > actually points to the end of the checkpoint record. But how come the min > recovery ending location's timeline is 1, while the checkpoint record's > timeline is 2. > > Now maybe that would happen to work if remove the sanity check, but it still > seems horribly confusing. I'm afraid that discrepancy will come back to > haunt us later if we leave it like that. So I'd like to fix that. > > Mulling over this for some more, I propose the attached patch. With the > patch, we peek into the checkpoint record, and actually perform the timeline > switch (by changing ThisTimeLineID) before replaying it. That way the > checkpoint record is really considered to be on the new timeline for all > purposes. At the moment, the only difference that makes in practice is that > we set replayEndTLI, and thus minRecoveryPointTLI, to the new TLI, but it > feels logically more correct to do it that way. This patch has already been included in HEAD. Right? I found another "requested timeline does not contain minimum recovery point" error scenario in HEAD: 1. Set up the master 'M', one standby 'S1', and one cascade standby 'S2'. 2. Shutdown the master 'M' and promote the standby 'S1', and wait for 'S2' to reconnect to 'S1'. 3. Set up new cascade standby 'S3' connecting to 'S2'. Then 'S3' fails to start the recovery because of the following error: FATAL: requested timeline 2 does not contain minimum recovery point 0/3000000 on timeline 1 LOG: startup process (PID 33104) exited with exit code 1 LOG: aborting startup due tostartup process failure The result of pg_controldata of 'S3' is: Latest checkpoint location: 0/3000088 Prior checkpoint location: 0/2000060 Latest checkpoint's REDO location: 0/3000088 Latest checkpoint's REDO WAL file: 000000020000000000000003 Latest checkpoint's TimeLineID: 2 <snip> Min recovery ending location: 0/3000000 Min recovery ending loc's timeline: 1 Backup start location: 0/0 Backup end location: 0/0 The content of the timeline history file '00000002.history' is: 1 0/3000088 no recovery target specified Regards, -- Fujii Masao
On 15.12.2012 01:09, Josh Berkus wrote: > Tested this on yesterday's snapshot. Worked great. Thanks for the testing! > Now I wanna test a chain of cascading replicas ... how far can we chain > these? There's no limit in theory. I tested with one master and two chained standbys myself. Give it a shot, I'm curious to hear how it works with a chain of a hundred standbys ;-). - Heikki
On 17 December 2012 12:07, Heikki Linnakangas <hlinnakangas@vmware.com> wrote:
-- On 15.12.2012 01:09, Josh Berkus wrote:Thanks for the testing!Tested this on yesterday's snapshot. Worked great.There's no limit in theory. I tested with one master and two chained standbys myself. Give it a shot, I'm curious to hear how it works with a chain of a hundred standbys ;-).Now I wanna test a chain of cascading replicas ... how far can we chain
these?
I just set up 120 chained standbys, and for some reason I'm seeing these errors:
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 1
LOG: record with zero length at 0/301EC10
LOG: fetching timeline history file for timeline 2 from primary server
LOG: restarted WAL streaming at 0/3000000 on timeline 1
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 1
LOG: new target timeline is 2
LOG: restarted WAL streaming at 0/3000000 on timeline 2
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 2
FATAL: error reading result of streaming command: ERROR: requested WAL segment 000000020000000000000003 has already been removed
ERROR: requested WAL segment 000000020000000000000003 has already been removed
LOG: started streaming WAL from primary at 0/3000000 on timeline 2
ERROR: requested WAL segment 000000020000000000000003 has already been removed
The "End of WAL reached on timeline 2" appears on all standbys except the one streaming directly from the primary.
However, changes continue to cascade to all standbys right to the end of the chain (it takes several minutes to propagate however).
Thom
Since Thom already did the destruction test, I only chained 7 standbies, just to see if I could reproduce his error. In the process, I accidentally connected one standby to itself. This failed, but the error message wasn't very helpful; it just gave me "FATAL: could not connect, the database system is starting up". Surely there's some way we could tell the user they've tried to connect a standby to itself? Anyway, I was unable to reproduce Thom's error. I did not see the error message he did. Without any read queries running on the standbys, lag from master to replica7 averaged about 0.5 seconds, ranging between 0.1 seconds and 1.2 seconds. -- Josh Berkus PostgreSQL Experts Inc. http://pgexperts.com
Heikki, I ran into an unexpected issue while testing. I just wanted to fire up a chain of 5 replicas to see if I could connect them in a loop. However, I ran into a weird issue when starting up "r3": it refused to come out of "the database is starting up" mode until I did a write on the master. Then it came up fine. master-->r1-->r2-->r3-->r4 I tried doing the full replication sequence (basebackup, startup, test) with it twice and got the exact same results each time. This is very strange because I did not encounter the same issues with r2 or r4. Nor have I seen this before in my tests. I'm also seeing Thom's spurious error message now. Each of r2, r3 and r4 have the following message once in their logs: LOG: database system was interrupted while in recovery at log time 2012-12-19 02:49:34 GMT HINT: If this has occurred more than once some data might be corrupted and you might need to choose an earlier recovery target. This message doesn't seem to signify anything. -- Josh Berkus PostgreSQL Experts Inc. http://pgexperts.com
On 19.12.2012 04:57, Josh Berkus wrote: > Heikki, > > I ran into an unexpected issue while testing. I just wanted to fire up > a chain of 5 replicas to see if I could connect them in a loop. > However, I ran into a weird issue when starting up "r3": it refused to > come out of "the database is starting up" mode until I did a write on > the master. Then it came up fine. > > master-->r1-->r2-->r3-->r4 > > I tried doing the full replication sequence (basebackup, startup, test) > with it twice and got the exact same results each time. > > This is very strange because I did not encounter the same issues with r2 > or r4. Nor have I seen this before in my tests. Ok.. I'm going to need some more details on how to reproduce this, I'm not seeing that when I set up four standbys. > I'm also seeing Thom's spurious error message now. Each of r2, r3 and > r4 have the following message once in their logs: > > LOG: database system was interrupted while in recovery at log time > 2012-12-19 02:49:34 GMT > HINT: If this has occurred more than once some data might be corrupted > and you might need to choose an earlier recovery target. > > This message doesn't seem to signify anything. Yep. You get that message when you start up the system from a base backup that was taken from a standby server. It's just noise, it would be nice if we could dial it down somehow. In general, streaming replication and backups tend to be awfully noisy. I've been meaning to go through all the messages that get printed during normal operation and think carefully which ones are really necessary, which ones could perhaps be merged into more compact messages. But haven't gotten around to it; that would be a great project for someone who actually sets up these systems regularly in production. - Heikki
On 19.12.2012 15:55, Heikki Linnakangas wrote: > On 19.12.2012 04:57, Josh Berkus wrote: >> Heikki, >> >> I ran into an unexpected issue while testing. I just wanted to fire up >> a chain of 5 replicas to see if I could connect them in a loop. >> However, I ran into a weird issue when starting up "r3": it refused to >> come out of "the database is starting up" mode until I did a write on >> the master. Then it came up fine. >> >> master-->r1-->r2-->r3-->r4 >> >> I tried doing the full replication sequence (basebackup, startup, test) >> with it twice and got the exact same results each time. >> >> This is very strange because I did not encounter the same issues with r2 >> or r4. Nor have I seen this before in my tests. > > Ok.. I'm going to need some more details on how to reproduce this, I'm > not seeing that when I set up four standbys. Ok, I managed to reproduce this now. The problem seems to be a timing problem, when a standby switches to follow a new timeline. Four is not a magic number here, it can happen with just one cascading standby too. When the timline switch happens, for example, the standby changes recovery target timeline from 1 to 2, at WAL position 0/30002D8, it has all the WAL up to that WAL position. However, it only has that WAL in file 000000010000000000000003, corresponding to timeline 1, and not in the file 000000020000000000000003, corresponding to the new timeline. When a cascaded standby connects, it requests to start streaming from point 0/3000000 at timeline 2 (we always start streaming from the beginning of a segment, to avoid leaving partially-filled segments in pg_xlog). The walsender in the 1st standby tries to read that from file 000000020000000000000003, which does not exist yet. The problem goes away after some time, after the 1st standby has streamed the contents of 000000020000000000000003 and written it to disk, and the cascaded standby reconnects. But it would be nice to avoid that situation. I'm not sure how to do that yet, we might need to track the timeline we're currently receiving/sending more carefully. Or perhaps we need to copy the previous WAL segment to the new name when switching recovery target timeline, like we do when a server is promoted. I'll try to come up with something... - Heikki
On 19.12.2012 17:27, Heikki Linnakangas wrote: > On 19.12.2012 15:55, Heikki Linnakangas wrote: >> On 19.12.2012 04:57, Josh Berkus wrote: >>> Heikki, >>> >>> I ran into an unexpected issue while testing. I just wanted to fire up >>> a chain of 5 replicas to see if I could connect them in a loop. >>> However, I ran into a weird issue when starting up "r3": it refused to >>> come out of "the database is starting up" mode until I did a write on >>> the master. Then it came up fine. >>> >>> master-->r1-->r2-->r3-->r4 >>> >>> I tried doing the full replication sequence (basebackup, startup, test) >>> with it twice and got the exact same results each time. >>> >>> This is very strange because I did not encounter the same issues with r2 >>> or r4. Nor have I seen this before in my tests. >> >> Ok.. I'm going to need some more details on how to reproduce this, I'm >> not seeing that when I set up four standbys. > > Ok, I managed to reproduce this now. Hmph, no I didn't, I replied to wrong email. The problem I managed to reproduce was the one where you get "requested WAL segment 000000020000000000000003 has already been removed" errors, reported by Thom. - Heikki
Heikki, > The problem goes away after some time, after the 1st standby has > streamed the contents of 000000020000000000000003 and written it to > disk, and the cascaded standby reconnects. But it would be nice to > avoid > that situation. I'm not sure how to do that yet, we might need to > track > the timeline we're currently receiving/sending more carefully. Or > perhaps we need to copy the previous WAL segment to the new name when > switching recovery target timeline, like we do when a server is > promoted. I'll try to come up with something... Would it be accurate to say that this issue only happens when all of the replicated servers have no traffic? --Josh
Heikki, The next time I get the issue, and I'm not paying for 5 cloud servers by the hour, I'll give you a login. --Josh ----- Original Message ----- > On 19.12.2012 17:27, Heikki Linnakangas wrote: > > On 19.12.2012 15:55, Heikki Linnakangas wrote: > >> On 19.12.2012 04:57, Josh Berkus wrote: > >>> Heikki, > >>> > >>> I ran into an unexpected issue while testing. I just wanted to > >>> fire up > >>> a chain of 5 replicas to see if I could connect them in a loop. > >>> However, I ran into a weird issue when starting up "r3": it > >>> refused to > >>> come out of "the database is starting up" mode until I did a > >>> write on > >>> the master. Then it came up fine. > >>> > >>> master-->r1-->r2-->r3-->r4 > >>> > >>> I tried doing the full replication sequence (basebackup, startup, > >>> test) > >>> with it twice and got the exact same results each time. > >>> > >>> This is very strange because I did not encounter the same issues > >>> with r2 > >>> or r4. Nor have I seen this before in my tests. > >> > >> Ok.. I'm going to need some more details on how to reproduce this, > >> I'm > >> not seeing that when I set up four standbys. > > > > Ok, I managed to reproduce this now. > > Hmph, no I didn't, I replied to wrong email. The problem I managed to > reproduce was the one where you get "requested WAL > segment 000000020000000000000003 has already been removed" errors, > reported by Thom. > > - Heikki >
On 17.12.2012 15:05, Thom Brown wrote: > I just set up 120 chained standbys, and for some reason I'm seeing these > errors: > > LOG: replication terminated by primary server > DETAIL: End of WAL reached on timeline 1 > LOG: record with zero length at 0/301EC10 > LOG: fetching timeline history file for timeline 2 from primary server > LOG: restarted WAL streaming at 0/3000000 on timeline 1 > LOG: replication terminated by primary server > DETAIL: End of WAL reached on timeline 1 > LOG: new target timeline is 2 > LOG: restarted WAL streaming at 0/3000000 on timeline 2 > LOG: replication terminated by primary server > DETAIL: End of WAL reached on timeline 2 > FATAL: error reading result of streaming command: ERROR: requested WAL > segment 000000020000000000000003 has already been removed > > ERROR: requested WAL segment 000000020000000000000003 has already been > removed > LOG: started streaming WAL from primary at 0/3000000 on timeline 2 > ERROR: requested WAL segment 000000020000000000000003 has already been > removed I just committed a patch that should make the "requested WAL segment 000000020000000000000003 has already been removed" errors go away. The trick was for walsenders to not switch to the new timeline until at least one record has been replayed on it. That closes the window where the walsender already considers the new timeline to be the latest, but the WAL file has not been created yet. - Heikki
On 2012-12-20 14:45:05 +0200, Heikki Linnakangas wrote: > On 17.12.2012 15:05, Thom Brown wrote: > >I just set up 120 chained standbys, and for some reason I'm seeing these > >errors: > > > >LOG: replication terminated by primary server > >DETAIL: End of WAL reached on timeline 1 > >LOG: record with zero length at 0/301EC10 > >LOG: fetching timeline history file for timeline 2 from primary server > >LOG: restarted WAL streaming at 0/3000000 on timeline 1 > >LOG: replication terminated by primary server > >DETAIL: End of WAL reached on timeline 1 > >LOG: new target timeline is 2 > >LOG: restarted WAL streaming at 0/3000000 on timeline 2 > >LOG: replication terminated by primary server > >DETAIL: End of WAL reached on timeline 2 > >FATAL: error reading result of streaming command: ERROR: requested WAL > >segment 000000020000000000000003 has already been removed > > > >ERROR: requested WAL segment 000000020000000000000003 has already been > >removed > >LOG: started streaming WAL from primary at 0/3000000 on timeline 2 > >ERROR: requested WAL segment 000000020000000000000003 has already been > >removed > > I just committed a patch that should make the "requested WAL segment > 000000020000000000000003 has already been removed" errors go away. The trick > was for walsenders to not switch to the new timeline until at least one > record has been replayed on it. That closes the window where the walsender > already considers the new timeline to be the latest, but the WAL file has > not been created yet. I vote for introducing InvalidTimeLineID soon... 0 as a invalid TimeLineID seems to spread and is annoying to grep for. Greetings, Andres Freund --Andres Freund http://www.2ndQuadrant.com/PostgreSQL Development, 24x7 Support, Training & Services
On Sat, Dec 15, 2012 at 9:36 AM, Fujii Masao <masao.fujii@gmail.com> wrote: > On Sat, Dec 8, 2012 at 12:51 AM, Heikki Linnakangas > <hlinnakangas@vmware.com> wrote: >> On 06.12.2012 15:39, Amit Kapila wrote: >>> >>> On Thursday, December 06, 2012 12:53 AM Heikki Linnakangas wrote: >>>> >>>> On 05.12.2012 14:32, Amit Kapila wrote: >>>>> >>>>> On Tuesday, December 04, 2012 10:01 PM Heikki Linnakangas wrote: >>>>>> >>>>>> After some diversions to fix bugs and refactor existing code, I've >>>>>> committed a couple of small parts of this patch, which just add some >>>>>> sanity checks to notice incorrect PITR scenarios. Here's a new >>>>>> version of the main patch based on current HEAD. >>>>> >>>>> >>>>> After testing with the new patch, the following problems are observed. >>>>> >>>>> Defect - 1: >>>>> >>>>> 1. start primary A >>>>> 2. start standby B following A >>>>> 3. start cascade standby C following B. >>>>> 4. start another standby D following C. >>>>> 5. Promote standby B. >>>>> 6. After successful time line switch in cascade standby C& D, >>>> >>>> stop D. >>>>> >>>>> 7. Restart D, Startup is successful and connecting to standby C. >>>>> 8. Stop C. >>>>> 9. Restart C, startup is failing. >>>> >>>> >>>> Ok, the error I get in that scenario is: >>>> >>>> C 2012-12-05 19:55:43.840 EET 9283 FATAL: requested timeline 2 does not >>>> contain minimum recovery point 0/3023F08 on timeline 1 C 2012-12-05 >>>> 19:55:43.841 EET 9282 LOG: startup process (PID 9283) exited with exit >>>> code 1 C 2012-12-05 19:55:43.841 EET 9282 LOG: aborting startup due to >>>> startup process failure >>>> >>> >>>> >>>> That mismatch causes the error. I'd like to fix this by always treating >>>> the checkpoint record to be part of the new timeline. That feels more >>>> correct. The most straightforward way to implement that would be to peek >>>> at the xlog record before updating replayEndRecPtr and replayEndTLI. If >>>> it's a checkpoint record that changes TLI, set replayEndTLI to the new >>>> timeline before calling the redo-function. But it's a bit of a >>>> modularity violation to peek into the record like that. >>>> >>>> Or we could just revert the sanity check at beginning of recovery that >>>> throws the "requested timeline 2 does not contain minimum recovery point >>>> 0/3023F08 on timeline 1" error. The error I added to redo of checkpoint >>>> record that says "unexpected timeline ID %u in checkpoint record, before >>>> reaching minimum recovery point %X/%X on timeline %u" checks basically >>>> the same thing, but at a later stage. However, the way >>>> minRecoveryPointTLI is updated still seems wrong to me, so I'd like to >>>> fix that. >>>> >>>> I'm thinking of something like the attached (with some more comments >>>> before committing). Thoughts? >>> >>> >>> This has fixed the problem reported. >>> However, I am not able to think will there be any problem if we remove >>> check >>> "requested timeline 2 does not contain minimum recovery point >>>> >>>> 0/3023F08 on timeline 1" at beginning of recovery and just update >>> >>> replayEndTLI with ThisTimeLineID? >> >> >> Well, it seems wrong for the control file to contain a situation like this: >> >> pg_control version number: 932 >> Catalog version number: 201211281 >> Database system identifier: 5819228770976387006 >> Database cluster state: shut down in recovery >> pg_control last modified: pe 7. joulukuuta 2012 17.39.57 >> Latest checkpoint location: 0/3023EA8 >> Prior checkpoint location: 0/2000060 >> Latest checkpoint's REDO location: 0/3023EA8 >> Latest checkpoint's REDO WAL file: 000000020000000000000003 >> Latest checkpoint's TimeLineID: 2 >> ... >> Time of latest checkpoint: pe 7. joulukuuta 2012 17.39.49 >> Min recovery ending location: 0/3023F08 >> Min recovery ending loc's timeline: 1 >> >> Note the latest checkpoint location and its TimelineID, and compare them >> with the min recovery ending location. The min recovery ending location is >> ahead of latest checkpoint's location; the min recovery ending location >> actually points to the end of the checkpoint record. But how come the min >> recovery ending location's timeline is 1, while the checkpoint record's >> timeline is 2. >> >> Now maybe that would happen to work if remove the sanity check, but it still >> seems horribly confusing. I'm afraid that discrepancy will come back to >> haunt us later if we leave it like that. So I'd like to fix that. >> >> Mulling over this for some more, I propose the attached patch. With the >> patch, we peek into the checkpoint record, and actually perform the timeline >> switch (by changing ThisTimeLineID) before replaying it. That way the >> checkpoint record is really considered to be on the new timeline for all >> purposes. At the moment, the only difference that makes in practice is that >> we set replayEndTLI, and thus minRecoveryPointTLI, to the new TLI, but it >> feels logically more correct to do it that way. > > This patch has already been included in HEAD. Right? > > I found another "requested timeline does not contain minimum recovery point" > error scenario in HEAD: > > 1. Set up the master 'M', one standby 'S1', and one cascade standby 'S2'. > 2. Shutdown the master 'M' and promote the standby 'S1', and wait for 'S2' > to reconnect to 'S1'. > 3. Set up new cascade standby 'S3' connecting to 'S2'. > Then 'S3' fails to start the recovery because of the following error: > > FATAL: requested timeline 2 does not contain minimum recovery > point 0/3000000 on timeline 1 > LOG: startup process (PID 33104) exited with exit code 1 > LOG: aborting startup due to startup process failure > > The result of pg_controldata of 'S3' is: > > Latest checkpoint location: 0/3000088 > Prior checkpoint location: 0/2000060 > Latest checkpoint's REDO location: 0/3000088 > Latest checkpoint's REDO WAL file: 000000020000000000000003 > Latest checkpoint's TimeLineID: 2 > <snip> > Min recovery ending location: 0/3000000 > Min recovery ending loc's timeline: 1 > Backup start location: 0/0 > Backup end location: 0/0 > > The content of the timeline history file '00000002.history' is: > > 1 0/3000088 no recovery target specified I still could reproduce this problem. Attached is the shell script which reproduces the problem. Regards, -- Fujii Masao
Attachment
> I just committed a patch that should make the "requested WAL segment > 000000020000000000000003 has already been removed" errors go away. > The > trick was for walsenders to not switch to the new timeline until at > least one record has been replayed on it. That closes the window > where > the walsender already considers the new timeline to be the latest, > but > the WAL file has not been created yet. OK, I'll download the snapshot in a couple days and make sure this didn't breaks something else. --Josh
On 20 December 2012 12:45, Heikki Linnakangas <hlinnakangas@vmware.com> wrote:
Thom On 17.12.2012 15:05, Thom Brown wrote:I just committed a patch that should make the "requested WAL segment 000000020000000000000003 has already been removed" errors go away. The trick was for walsenders to not switch to the new timeline until at least one record has been replayed on it. That closes the window where the walsender already considers the new timeline to be the latest, but the WAL file has not been created yet.I just set up 120 chained standbys, and for some reason I'm seeing these
errors:
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 1
LOG: record with zero length at 0/301EC10
LOG: fetching timeline history file for timeline 2 from primary server
LOG: restarted WAL streaming at 0/3000000 on timeline 1
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 1
LOG: new target timeline is 2
LOG: restarted WAL streaming at 0/3000000 on timeline 2
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 2
FATAL: error reading result of streaming command: ERROR: requested WAL
segment 000000020000000000000003 has already been removed
ERROR: requested WAL segment 000000020000000000000003 has already been
removed
LOG: started streaming WAL from primary at 0/3000000 on timeline 2
ERROR: requested WAL segment 000000020000000000000003 has already been
removed
Now I'm getting this on all standbys after promoting the first standby in a chain.
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 1
LOG: record with zero length at 0/301EC10
LOG: fetching timeline history file for timeline 2 from primary server
LOG: restarted WAL streaming at 0/3000000 on timeline 1
FATAL: could not receive data from WAL stream:
LOG: new target timeline is 2
FATAL: could not connect to the primary server: FATAL: the database system is in recovery mode
LOG: started streaming WAL from primary at 0/3000000 on timeline 2
TRAP: FailedAssertion("!(((sentPtr) <= (SendRqstPtr)))", File: "walsender.c", Line: 1425)
LOG: server process (PID 19917) was terminated by signal 6: Aborted
LOG: terminating any other active server processes
LOG: all server processes terminated; reinitializing
LOG: database system was interrupted while in recovery at log time 2012-12-20 23:41:23 GMT
HINT: If this has occurred more than once some data might be corrupted and you might need to choose an earlier recovery target.
LOG: entering standby mode
FATAL: the database system is in recovery mode
LOG: redo starts at 0/2000028
LOG: consistent recovery state reached at 0/20000E8
LOG: database system is ready to accept read only connections
LOG: record with zero length at 0/301EC70
LOG: started streaming WAL from primary at 0/3000000 on timeline 2
LOG: unexpected EOF on standby connection
And if I restart the new primary, the first new standby connected to it shows:
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 2
FATAL: error reading result of streaming command: server closed the connection unexpectedly
This probably means the server terminated abnormally
before or while processing the request.
LOG: record with zero length at 0/301F1E0
However, all other standbys don't show any additional log output.
--
On 21.12.2012 01:50, Thom Brown wrote: > Now I'm getting this on all standbys after promoting the first standby in a > chain.> ...> TRAP: FailedAssertion("!(((sentPtr)<= (SendRqstPtr)))", File:> "walsender.c", Line: 1425) Sigh. I'm sounding like a broken record, but I just committed another fix for this, should work now. - Heikki
On 21 December 2012 18:13, Heikki Linnakangas <hlinnakangas@vmware.com> wrote:
-- On 21.12.2012 01:50, Thom Brown wrote:> ...Now I'm getting this on all standbys after promoting the first standby in a
chain.Sigh. I'm sounding like a broken record, but I just committed another fix for this, should work now.
> TRAP: FailedAssertion("!(((sentPtr)<= (SendRqstPtr)))", File:
> "walsender.c", Line: 1425)
Thanks Heikki. Just quickly retested with a new set of 120 standbys and all looks fine as far as the logs are concerned:
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 1
LOG: record with zero length at 0/37902A0
LOG: fetching timeline history file for timeline 2 from primary server
LOG: restarted WAL streaming at 0/3000000 on timeline 1
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 1
LOG: new target timeline is 2
LOG: restarted WAL streaming at 0/3000000 on timeline 2
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 2
LOG: record with zero length at 0/643E248
LOG: fetching timeline history file for timeline 3 from primary server
LOG: restarted WAL streaming at 0/6000000 on timeline 2
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 2
LOG: new target timeline is 3
LOG: restarted WAL streaming at 0/6000000 on timeline 3
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 3
LOG: record with zero length at 0/6BB13A8
LOG: fetching timeline history file for timeline 4 from primary server
LOG: restarted WAL streaming at 0/6000000 on timeline 3
LOG: replication terminated by primary server
DETAIL: End of WAL reached on timeline 3
LOG: new target timeline is 4
LOG: restarted WAL streaming at 0/6000000 on timeline 4
Thom
On Fri, Dec 21, 2012 at 1:48 AM, Fujii Masao <masao.fujii@gmail.com> wrote: > On Sat, Dec 15, 2012 at 9:36 AM, Fujii Masao <masao.fujii@gmail.com> wrote: >> On Sat, Dec 8, 2012 at 12:51 AM, Heikki Linnakangas >> <hlinnakangas@vmware.com> wrote: >>> On 06.12.2012 15:39, Amit Kapila wrote: >>>> >>>> On Thursday, December 06, 2012 12:53 AM Heikki Linnakangas wrote: >>>>> >>>>> On 05.12.2012 14:32, Amit Kapila wrote: >>>>>> >>>>>> On Tuesday, December 04, 2012 10:01 PM Heikki Linnakangas wrote: >>>>>>> >>>>>>> After some diversions to fix bugs and refactor existing code, I've >>>>>>> committed a couple of small parts of this patch, which just add some >>>>>>> sanity checks to notice incorrect PITR scenarios. Here's a new >>>>>>> version of the main patch based on current HEAD. >>>>>> >>>>>> >>>>>> After testing with the new patch, the following problems are observed. >>>>>> >>>>>> Defect - 1: >>>>>> >>>>>> 1. start primary A >>>>>> 2. start standby B following A >>>>>> 3. start cascade standby C following B. >>>>>> 4. start another standby D following C. >>>>>> 5. Promote standby B. >>>>>> 6. After successful time line switch in cascade standby C& D, >>>>> >>>>> stop D. >>>>>> >>>>>> 7. Restart D, Startup is successful and connecting to standby C. >>>>>> 8. Stop C. >>>>>> 9. Restart C, startup is failing. >>>>> >>>>> >>>>> Ok, the error I get in that scenario is: >>>>> >>>>> C 2012-12-05 19:55:43.840 EET 9283 FATAL: requested timeline 2 does not >>>>> contain minimum recovery point 0/3023F08 on timeline 1 C 2012-12-05 >>>>> 19:55:43.841 EET 9282 LOG: startup process (PID 9283) exited with exit >>>>> code 1 C 2012-12-05 19:55:43.841 EET 9282 LOG: aborting startup due to >>>>> startup process failure >>>>> >>>> >>>>> >>>>> That mismatch causes the error. I'd like to fix this by always treating >>>>> the checkpoint record to be part of the new timeline. That feels more >>>>> correct. The most straightforward way to implement that would be to peek >>>>> at the xlog record before updating replayEndRecPtr and replayEndTLI. If >>>>> it's a checkpoint record that changes TLI, set replayEndTLI to the new >>>>> timeline before calling the redo-function. But it's a bit of a >>>>> modularity violation to peek into the record like that. >>>>> >>>>> Or we could just revert the sanity check at beginning of recovery that >>>>> throws the "requested timeline 2 does not contain minimum recovery point >>>>> 0/3023F08 on timeline 1" error. The error I added to redo of checkpoint >>>>> record that says "unexpected timeline ID %u in checkpoint record, before >>>>> reaching minimum recovery point %X/%X on timeline %u" checks basically >>>>> the same thing, but at a later stage. However, the way >>>>> minRecoveryPointTLI is updated still seems wrong to me, so I'd like to >>>>> fix that. >>>>> >>>>> I'm thinking of something like the attached (with some more comments >>>>> before committing). Thoughts? >>>> >>>> >>>> This has fixed the problem reported. >>>> However, I am not able to think will there be any problem if we remove >>>> check >>>> "requested timeline 2 does not contain minimum recovery point >>>>> >>>>> 0/3023F08 on timeline 1" at beginning of recovery and just update >>>> >>>> replayEndTLI with ThisTimeLineID? >>> >>> >>> Well, it seems wrong for the control file to contain a situation like this: >>> >>> pg_control version number: 932 >>> Catalog version number: 201211281 >>> Database system identifier: 5819228770976387006 >>> Database cluster state: shut down in recovery >>> pg_control last modified: pe 7. joulukuuta 2012 17.39.57 >>> Latest checkpoint location: 0/3023EA8 >>> Prior checkpoint location: 0/2000060 >>> Latest checkpoint's REDO location: 0/3023EA8 >>> Latest checkpoint's REDO WAL file: 000000020000000000000003 >>> Latest checkpoint's TimeLineID: 2 >>> ... >>> Time of latest checkpoint: pe 7. joulukuuta 2012 17.39.49 >>> Min recovery ending location: 0/3023F08 >>> Min recovery ending loc's timeline: 1 >>> >>> Note the latest checkpoint location and its TimelineID, and compare them >>> with the min recovery ending location. The min recovery ending location is >>> ahead of latest checkpoint's location; the min recovery ending location >>> actually points to the end of the checkpoint record. But how come the min >>> recovery ending location's timeline is 1, while the checkpoint record's >>> timeline is 2. >>> >>> Now maybe that would happen to work if remove the sanity check, but it still >>> seems horribly confusing. I'm afraid that discrepancy will come back to >>> haunt us later if we leave it like that. So I'd like to fix that. >>> >>> Mulling over this for some more, I propose the attached patch. With the >>> patch, we peek into the checkpoint record, and actually perform the timeline >>> switch (by changing ThisTimeLineID) before replaying it. That way the >>> checkpoint record is really considered to be on the new timeline for all >>> purposes. At the moment, the only difference that makes in practice is that >>> we set replayEndTLI, and thus minRecoveryPointTLI, to the new TLI, but it >>> feels logically more correct to do it that way. >> >> This patch has already been included in HEAD. Right? >> >> I found another "requested timeline does not contain minimum recovery point" >> error scenario in HEAD: >> >> 1. Set up the master 'M', one standby 'S1', and one cascade standby 'S2'. >> 2. Shutdown the master 'M' and promote the standby 'S1', and wait for 'S2' >> to reconnect to 'S1'. >> 3. Set up new cascade standby 'S3' connecting to 'S2'. >> Then 'S3' fails to start the recovery because of the following error: >> >> FATAL: requested timeline 2 does not contain minimum recovery >> point 0/3000000 on timeline 1 >> LOG: startup process (PID 33104) exited with exit code 1 >> LOG: aborting startup due to startup process failure >> >> The result of pg_controldata of 'S3' is: >> >> Latest checkpoint location: 0/3000088 >> Prior checkpoint location: 0/2000060 >> Latest checkpoint's REDO location: 0/3000088 >> Latest checkpoint's REDO WAL file: 000000020000000000000003 >> Latest checkpoint's TimeLineID: 2 >> <snip> >> Min recovery ending location: 0/3000000 >> Min recovery ending loc's timeline: 1 >> Backup start location: 0/0 >> Backup end location: 0/0 >> >> The content of the timeline history file '00000002.history' is: >> >> 1 0/3000088 no recovery target specified > > I still could reproduce this problem. Attached is the shell script > which reproduces the problem. This problem happens when new standby starts up from the backup taken from another standby and its recovery starts from the shutdown checkpoint record which causes timeline switch. In this case, the timeline of minimum recovery point can be different from that of latest checkpoint (i.e., shutdown checkpoint). But the following check in StartupXLOG() assumes that they are always the same wrongly. So the problem happens. /* * The min recovery point should be part of the requested timeline's * history, too. */if (!XLogRecPtrIsInvalid(ControlFile->minRecoveryPoint)&& tliOfPointInHistory(ControlFile->minRecoveryPoint - 1, expectedTLEs)!= ControlFile->minRecoveryPointTLI) ereport(FATAL, (errmsg("requested timeline %u doesnot contain minimum recovery point %X/%X on timeline %u", recoveryTargetTLI, (uint32) (ControlFile->minRecoveryPoint>> 32), (uint32) ControlFile->minRecoveryPoint, ControlFile->minRecoveryPointTLI))); If we don't have such check, in the later phase of recovery, the minimum recovery point is initialized to the latest checkpoint location as follows. This strikes me that the timeline of minimum recovery point should be check after it's initialized. So ISTM that the right fix of the problem is to move the above check after the following initialization. Thought? /* initialize minRecoveryPoint if not set yet */ if (XLByteLT(ControlFile->minRecoveryPoint, checkPoint.redo)) { ControlFile->minRecoveryPoint = checkPoint.redo; ControlFile->minRecoveryPointTLI= checkPoint.ThisTimeLineID; } Regards, -- Fujii Masao
On 23.12.2012 16:37, Fujii Masao wrote: > On Fri, Dec 21, 2012 at 1:48 AM, Fujii Masao<masao.fujii@gmail.com> wrote: >> On Sat, Dec 15, 2012 at 9:36 AM, Fujii Masao<masao.fujii@gmail.com> wrote: >>> I found another "requested timeline does not contain minimum recovery point" >>> error scenario in HEAD: >>> >>> 1. Set up the master 'M', one standby 'S1', and one cascade standby 'S2'. >>> 2. Shutdown the master 'M' and promote the standby 'S1', and wait for 'S2' >>> to reconnect to 'S1'. >>> 3. Set up new cascade standby 'S3' connecting to 'S2'. >>> Then 'S3' fails to start the recovery because of the following error: >>> >>> FATAL: requested timeline 2 does not contain minimum recovery >>> point 0/3000000 on timeline 1 >>> LOG: startup process (PID 33104) exited with exit code 1 >>> LOG: aborting startup due to startup process failure >>> >>> The result of pg_controldata of 'S3' is: >>> >>> Latest checkpoint location: 0/3000088 >>> Prior checkpoint location: 0/2000060 >>> Latest checkpoint's REDO location: 0/3000088 >>> Latest checkpoint's REDO WAL file: 000000020000000000000003 >>> Latest checkpoint's TimeLineID: 2 >>> <snip> >>> Min recovery ending location: 0/3000000 >>> Min recovery ending loc's timeline: 1 >>> Backup start location: 0/0 >>> Backup end location: 0/0 >>> >>> The content of the timeline history file '00000002.history' is: >>> >>> 1 0/3000088 no recovery target specified >> >> I still could reproduce this problem. Attached is the shell script >> which reproduces the problem. > > This problem happens when new standby starts up from the backup > taken from another standby and its recovery starts from the shutdown > checkpoint record which causes timeline switch. In this case, > the timeline of minimum recovery point can be different from that of > latest checkpoint (i.e., shutdown checkpoint). But the following check > in StartupXLOG() assumes that they are always the same wrongly. > So the problem happens. > > /* > * The min recovery point should be part of the requested timeline's > * history, too. > */ > if (!XLogRecPtrIsInvalid(ControlFile->minRecoveryPoint)&& > tliOfPointInHistory(ControlFile->minRecoveryPoint - 1, expectedTLEs) != > ControlFile->minRecoveryPointTLI) > ereport(FATAL, > (errmsg("requested timeline %u does not contain minimum recovery > point %X/%X on timeline %u", > recoveryTargetTLI, > (uint32) (ControlFile->minRecoveryPoint>> 32), > (uint32) ControlFile->minRecoveryPoint, > ControlFile->minRecoveryPointTLI))); No, it doesn't assume that min recovery point is on the same timeline as the checkpoint record. This is another variant of the "timeline history files are not included in the backup" problem discussed on the other thread with subject "pg_basebackup from cascading standby after timeline switch". If you remove the min recovery point check above, the test case still fails, with a different error message: LOG: unexpected timeline ID 1 in log segment 000000020000000000000003, offset 0 If you modify the test script to copy the 00000002.history file to the data-standby3/pg_xlog after running pg_basebackup, the test case works. (we still need to fix it, of course) - Heikki