Re: Reduce the time required for a database recovery from archive. - Mailing list pgsql-hackers
From | Dmitry Shulga |
---|---|
Subject | Re: Reduce the time required for a database recovery from archive. |
Date | |
Msg-id | 714C57C4-20B2-48C9-BDF9-C279183262C9@postgrespro.ru Whole thread Raw |
In response to | Re: Reduce the time required for a database recovery from archive. (Stephen Frost <sfrost@snowman.net>) |
Responses |
Re: Reduce the time required for a database recovery from archive.
|
List | pgsql-hackers |
Hello Stephen, > On 19 Oct 2020, at 23:25, Stephen Frost <sfrost@snowman.net> wrote: > > Greetings, > >>>> Implementation of this approach assumes running several background processes (bgworkers) >>>> each of which runs a shell command specified by the parameter restore_command >>>> to deliver an archived WAL file. Number of running parallel processes is limited >>>> by the new parameter max_restore_command_workers. If this parameter has value 0 >>>> then WAL files delivery is performed using the original algorithm, that is in >>>> one-by-one manner. If this parameter has value greater than 0 then the database >>>> server starts several bgworker processes up to the limit specified by >>>> the parameter max_restore_command_workers and passes to every process >>>> WAL file name to deliver. Active processes start prefetching of specified >>>> WAL files and store received files in the directory pg_wal/pgsql_tmp. After >>>> bgworker process finishes receiving a file it marks itself as a free process >>>> and waits for a new request to receive a next WAL file. The main process >>>> performing database recovery still handles WAL files in one-by-one manner, >>>> but instead of waiting for a next required WAL file's availability it checks for >>>> that file in the prefetched directory. If a new file is present there, >>>> the main process starts its processing. >>> >>> I'm a bit confused about this description- surely it makes sense for the >> OK. The description I originally provided was probably pretty misleading so I will try to clarify it a bit. >> >> So, as soon as a bgworker process finishes delivering a WAL file it marks itself as a free. >> >> WAL records applier working in parallel and processing the WAL files in sequential manner. >> Once it finishes handling of the current WAL file, it checks whether it is possible to run extra bgworker processes >> to deliver WAL files which will be required a bit later. If there are free bgworker processes then applier requests >> to start downloading of one or more extra WAL files. After that applier determines a name of next WAL file to handle >> and checks whether it exist in the prefetching directory. If it does exist then applier starts handling it and >> processing loop is repeated. > > Ok- so the idea is that each time the applying process finishes with a > WAL file then it'll see if there's an available worker and, if so, will > give it the next file to go get (which would presumably be some number > in the future and the actual next file the applying process needs is > already available). That sounds better, at least, though I'm not sure > why we're making it the job of the applying process to push the workers > each time..? Every bgwork serves as a task to deliver a WAL file. Considering a task as an active entity is well-known approach in softwaredesign. So I don't see any issues with such implementation. Moreover, implementation of this approach is probably simpler than anyother alternatives and still providing positive performance impact in comparing with current (non optimized) implementation. > Also, I'm not sure about the interface- wouldn't it make > more sense to have a "pre-fetch this amount of WAL" kind of parameter > directly instead of tying that to the number of background workers? This approach was originally considered and closely discussed. Finally, it was decided that introducing an extra GUC parameter to control pre-fetch limit is not practical since it shiftsresponsibility for tuning prefetching mechanism from postgres server to a user. From my point of view the fewer parameters exist to set up some feature the better. > You > might only need one or two processes doing WAL fetching to be able to > fetch faster than the applying process is able to apply it, but you > probably want to pre-fetch more than just one or two 16 MB WAL files. Every time when prefetching is started a number of potentially prefetched files is calculated by expression PREFETCH_RATION * max_restore_command_workers - 'number of already prefetched files' where PREFETCH_RATION is compiled-in constant and has value 16. After that a task for delivering a next WAL file is placed to a current free bgworker process up until no more free bgworkerprocesses. > In other words, I would have thought we'd have: > > wal_prefetch_amount = 1GB > max_restore_command_workers = 2 > > and then you'd have up to 2 worker processes running and they'd be > keeping 1GB of WAL pre-fetched at all times. If we have just > 'max_restore_command_workers' and you want to pre-fetch 1GB of WAL then > you'd have to have a pretty high value there and you'd end up with > a bunch of threads that all spike to go do work each time the applying Sorry, I don't see how we can end up with a bunch of threads? max_restore_command_workers has value 2 in your example meaning that no more than 2 bgworkers could be run concurrently forthe sake of WAL files prefetching > process finishes a WAL file but then just sit around doing nothing while > waiting for the applying process to finish another segment. I believe that for typical set-up the parameter max_restore_command_workers would have value 2 or 3 in order to supply a delivered WAL file on time just before it be started processing. This use case is for environment where time required for delivering WAL file from archive is greater than time requiredfor applying records contained in the WAL file. If time required for WAL file delivering lesser than than time required for handling records contained in it then max_restore_command_workersshouldn't be specified at all > > Thanks, > > Stephen
pgsql-hackers by date: