Thread: Full page writes improvement, code update

Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

Here's an update of a code to improve full page writes as proposed in

http://archives.postgresql.org/pgsql-hackers/2007-01/msg01491.php
and
http://archives.postgresql.org/pgsql-patches/2007-01/msg00607.php

Update includes some modification for error handling in archiver and
restoration command.

In the previous threads, I posted several evaluation and shown that we
can keep all the full page writes needed for full XLOG crash recovery,
while compressing archive log size considerably better than gzip, with
less CPU consumption.  I've found no further objection for this proposal
but still would like to hear comments/opinions/advices.

Regards;

--
Koichi Suzuki

Attachment

Re: Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Tue, 2007-03-27 at 11:52 +0900, Koichi Suzuki wrote:

> Here's an update of a code to improve full page writes as proposed in
>
> http://archives.postgresql.org/pgsql-hackers/2007-01/msg01491.php
> and
> http://archives.postgresql.org/pgsql-patches/2007-01/msg00607.php
>
> Update includes some modification for error handling in archiver and
> restoration command.
>
> In the previous threads, I posted several evaluation and shown that we
> can keep all the full page writes needed for full XLOG crash recovery,
> while compressing archive log size considerably better than gzip, with
> less CPU consumption.  I've found no further objection for this proposal
> but still would like to hear comments/opinions/advices.

Koichi-san,

Looks interesting. I like the small amount of code to do this.

A few thoughts:

- Not sure why we need "full_page_compress", why not just mark them
always? That harms noone. (Did someone else ask for that? If so, keep
it)

- OTOH I'd like to see an explicit parameter set during recovery since
you're asking the main recovery path to act differently in case a single
bit is set/unset. If you are using that form of recovery, we should say
so explicitly, to keep everybody else safe.

- I'd rather mark just the nonremovable blocks. But no real difference

- We definitely don't want an normal elog in a XLogInsert critical
section, especially at DEBUG1 level

- diff -c format is easier and the standard

- pg_logarchive and pg_logrestore should be called by a name that
reflects what they actually do. Possibly pg_compresslog and
pg_decompresslog etc.. I've not reviewed those programs, but:

- Not sure why we have to compress away page headers. Touch as little as
you can has always been my thinking with recovery code.

- I'm very uncomfortable with touching the LSN. Maybe I misunderstand?

- Have you thought about how pg_standby would integrate with this
option? Can you please?

- I'll do some docs for this after Freeze, if you'd like. I have some
other changes to make there, so I can do this at the same time.

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Simon;

Thanks a lot for your comments/advices. I'd like to write some feedback.

Simon Riggs wrote:
> On Tue, 2007-03-27 at 11:52 +0900, Koichi Suzuki wrote:
>
>> Here's an update of a code to improve full page writes as proposed in
>>
>> http://archives.postgresql.org/pgsql-hackers/2007-01/msg01491.php
>> and
>> http://archives.postgresql.org/pgsql-patches/2007-01/msg00607.php
>>
>> Update includes some modification for error handling in archiver and
>> restoration command.
>>
>> In the previous threads, I posted several evaluation and shown that we
>> can keep all the full page writes needed for full XLOG crash recovery,
>> while compressing archive log size considerably better than gzip, with
>> less CPU consumption.  I've found no further objection for this proposal
>> but still would like to hear comments/opinions/advices.
>
> Koichi-san,
>
> Looks interesting. I like the small amount of code to do this.
>
> A few thoughts:
>
> - Not sure why we need "full_page_compress", why not just mark them
> always? That harms noone. (Did someone else ask for that? If so, keep
> it)

No, no one asked to have a separate option. There'll be no bad
influence to do so. As written below, full page write can be
categolized as follows:

1) Needed for crash recovery: first page update after each checkpoint.
This has to be kept in WAL.

2) Needed for archive recovery: page update between pg_start_backup and
pg_stop_backup. This has to be kept in archive log.

3) For log-shipping slave such as pg_standby: no full page writes will
be needed for this purpose.

My proposal deals with 2). So, if we mark each "full_page_write", I'd
rather mark when this is needed. Still need only one bit because the
case 3) does not need any mark.

> - OTOH I'd like to see an explicit parameter set during recovery since
> you're asking the main recovery path to act differently in case a single
> bit is set/unset. If you are using that form of recovery, we should say
> so explicitly, to keep everybody else safe.

Only one thing I had to do is to create "dummy" full page write to
maintain LSNs. Full page writes are omitted in archive log. We have to
LSNs same as those in the original WAL. In this case, recovery has to
read logical log, not "dummy" full page writes. On the other hand, if
both logical log and "real" full page writes are found in a log record,
the recovery has to use "real" full page writes.

> - I'd rather mark just the nonremovable blocks. But no real difference

It sound nicer. According to the full page write categories above, we
can mark full page writes as needed in "crash recovery" or "archive
recovery".   Please give some feedback to the above full page write
categories.

>
> - We definitely don't want an normal elog in a XLogInsert critical
> section, especially at DEBUG1 level

I agree. I'll remove elog calls from critical sections.

> - diff -c format is easier and the standard

I'll change the diff option.

> - pg_logarchive and pg_logrestore should be called by a name that
> reflects what they actually do. Possibly pg_compresslog and
> pg_decompresslog etc.. I've not reviewed those programs, but:

I wasn't careful to have command names more based on what to be done.
I'll change the command name.

>
> - Not sure why we have to compress away page headers. Touch as little as
> you can has always been my thinking with recovery code.

Eliminating page headers gives compression rate slightly better, a
couple of percents.   To make code simpler, I'll drop this compression.

>
> - I'm very uncomfortable with touching the LSN. Maybe I misunderstand?

The reason why we need pg_logarchive (or pg_decompresslog) is to
maintain LSN the same as those in the original WAL. For this purpose,
we need "dummy" full page write. I don't like to touch LSN either and
this is the reason why pg_decompresslog need some extra work,
eliminating many other changes in the recovery.

>
> - Have you thought about how pg_standby would integrate with this
> option? Can you please?

Yes I believe so. As pg_standby does not include any chance to meet
partial writes of pages, I believe you can omit all the full page
writes. Of course, as Tom Lange suggested in
http://archives.postgresql.org/pgsql-hackers/2007-02/msg00034.php
removing full page writes can lose a chance to recover from
partial/inconsisitent writes in the crash of pg_standby. In this case,
we have to import a backup and archive logs (with full page writes
during the backup) to recover. (We have to import them when the file
system crashes anyway). If it's okay, I believe
pg_compresslog/pg_decompresslog can be integrated with pg_standby.

Maybe we can work together to include pg_compresslog/pg_decompresslog in
pg_standby.

>
> - I'll do some docs for this after Freeze, if you'd like. I have some
> other changes to make there, so I can do this at the same time.

I'll be looking forward to your writings.

Best regards;

--
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Wed, 2007-03-28 at 10:54 +0900, Koichi Suzuki wrote:

> As written below, full page write can be
> categolized as follows:
>
> 1) Needed for crash recovery: first page update after each checkpoint.
> This has to be kept in WAL.
>
> 2) Needed for archive recovery: page update between pg_start_backup and
> pg_stop_backup. This has to be kept in archive log.
>
> 3) For log-shipping slave such as pg_standby: no full page writes will
> be needed for this purpose.
>
> My proposal deals with 2). So, if we mark each "full_page_write", I'd
> rather mark when this is needed. Still need only one bit because the
> case 3) does not need any mark.

I'm very happy with this proposal, though I do still have some points in
detailed areas.

If you accept that 1 & 2 are valid goals, then 1 & 3 or 1, 2 & 3 are
also valid goals, ISTM. i.e. you might choose to use full_page_writes on
the primary and yet would like to see optimised data transfer to the
standby server. In that case, you would need the mark.

> > - Not sure why we need "full_page_compress", why not just mark them
> > always? That harms noone. (Did someone else ask for that? If so, keep
> > it)
>
> No, no one asked to have a separate option. There'll be no bad
> influence to do so.  So, if we mark each "full_page_write", I'd
> rather mark when this is needed. Still need only one bit because the
> case 3) does not need any mark.

OK, different question:
Why would anyone ever set full_page_compress = off?

Why have a parameter that does so little? ISTM this is:

i) one more thing to get wrong

ii) cheaper to mark the block when appropriate than to perform the if()
test each time. That can be done only in the path where backup blocks
are present.

iii) If we mark the blocks every time, it allows us to do an offline WAL
compression. If the blocks aren't marked that option is lost. The bit is
useful information, so we should have it in all cases.

> > - OTOH I'd like to see an explicit parameter set during recovery since
> > you're asking the main recovery path to act differently in case a single
> > bit is set/unset. If you are using that form of recovery, we should say
> > so explicitly, to keep everybody else safe.
>
> Only one thing I had to do is to create "dummy" full page write to
> maintain LSNs. Full page writes are omitted in archive log. We have to
> LSNs same as those in the original WAL. In this case, recovery has to
> read logical log, not "dummy" full page writes. On the other hand, if
> both logical log and "real" full page writes are found in a log record,
> the recovery has to use "real" full page writes.

I apologise for not understanding your reply, perhaps my original
request was unclear.

In recovery.conf, I'd like to see a parameter such as

dummy_backup_blocks = off (default) | on

to explicitly indicate to the recovery process that backup blocks are
present, yet they are garbage and should be ignored. Having garbage data
within the system is potentially dangerous and I want to be told by the
user that they were expecting that and its OK to ignore that data.
Otherwise I want to throw informative errors. Maybe it seems OK now, but
the next change to the system may have unintended consequences and it
may not be us making the change. "It's OK the Alien will never escape
from the lab" is the starting premise for many good sci-fi horrors and I
want to watch them, not be in one myself. :-)

We can call it other things, of course. e.g.
ignore_dummy_blocks
decompressed_blocks
apply_backup_blocks

> Yes I believe so. As pg_standby does not include any chance to meet
> partial writes of pages, I believe you can omit all the full page
> writes. Of course, as Tom Lange suggested in
> http://archives.postgresql.org/pgsql-hackers/2007-02/msg00034.php
> removing full page writes can lose a chance to recover from
> partial/inconsisitent writes in the crash of pg_standby. In this case,
> we have to import a backup and archive logs (with full page writes
> during the backup) to recover. (We have to import them when the file
> system crashes anyway). If it's okay, I believe
> pg_compresslog/pg_decompresslog can be integrated with pg_standby.
>
> Maybe we can work together to include pg_compresslog/pg_decompresslog in
> pg_standby.

ISTM there are two options.

I think this option is already possible:

1. Allow pg_decompresslog to operate on a file, replacing it with the
expanded form, like gunzip, so we would do this:
  restore_command = 'pg_standby %f decomp.tmp && pg_decompresslog
decomp.tmp %p'

though the decomp.tmp file would not get properly initialised or cleaned
up when we finish.

whereas this will take additional work

2. Allow pg_standby to write to stdin, so that we can do this:
  restore_command = 'pg_standby %f | pg_decompresslog - %p'

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi, Here're some feedback to the comment:

Simon Riggs wrote:
> On Wed, 2007-03-28 at 10:54 +0900, Koichi Suzuki wrote:
>
>> As written below, full page write can be
>> categolized as follows:
>>
>> 1) Needed for crash recovery: first page update after each checkpoint.
>> This has to be kept in WAL.
>>
>> 2) Needed for archive recovery: page update between pg_start_backup and
>> pg_stop_backup. This has to be kept in archive log.
>>
>> 3) For log-shipping slave such as pg_standby: no full page writes will
>> be needed for this purpose.
>>
>> My proposal deals with 2). So, if we mark each "full_page_write", I'd
>> rather mark when this is needed. Still need only one bit because the
>> case 3) does not need any mark.
>
> I'm very happy with this proposal, though I do still have some points in
> detailed areas.
>
> If you accept that 1 & 2 are valid goals, then 1 & 3 or 1, 2 & 3 are
> also valid goals, ISTM. i.e. you might choose to use full_page_writes on
> the primary and yet would like to see optimised data transfer to the
> standby server. In that case, you would need the mark.

Yes, I need the mark.  In my proposal, only unmarked full-page-writes,
which were written as the first update after a checkpoint, are to be
removed offline (pg_compresslog).

>
>>> - Not sure why we need "full_page_compress", why not just mark them
>>> always? That harms noone. (Did someone else ask for that? If so, keep
>>> it)
>> No, no one asked to have a separate option. There'll be no bad
>> influence to do so.  So, if we mark each "full_page_write", I'd
>> rather mark when this is needed. Still need only one bit because the
>> case 3) does not need any mark.
>
> OK, different question:
> Why would anyone ever set full_page_compress = off?
>
> Why have a parameter that does so little? ISTM this is:
>
> i) one more thing to get wrong
>
> ii) cheaper to mark the block when appropriate than to perform the if()
> test each time. That can be done only in the path where backup blocks
> are present.
>
> iii) If we mark the blocks every time, it allows us to do an offline WAL
> compression. If the blocks aren't marked that option is lost. The bit is
> useful information, so we should have it in all cases.

Not only full-page-writes are written as WAL record.   In my proposal,
both full-page-writes and logical log are written in a WAL record, which
will make WAL size slightly bigger (five percent or so).   If
full_page_compress = off, only a full-page-write will be written in a
WAL record.   I thought someone will not be happy with this size growth.

I agree to make this mandatory if every body is happy with extra logical
  log in WAL records with full page writes.

I'd like to have your opinion.

>
>>> - OTOH I'd like to see an explicit parameter set during recovery since
>>> you're asking the main recovery path to act differently in case a single
>>> bit is set/unset. If you are using that form of recovery, we should say
>>> so explicitly, to keep everybody else safe.
>> Only one thing I had to do is to create "dummy" full page write to
>> maintain LSNs. Full page writes are omitted in archive log. We have to
>> LSNs same as those in the original WAL. In this case, recovery has to
>> read logical log, not "dummy" full page writes. On the other hand, if
>> both logical log and "real" full page writes are found in a log record,
>> the recovery has to use "real" full page writes.
>
> I apologise for not understanding your reply, perhaps my original
> request was unclear.
>
> In recovery.conf, I'd like to see a parameter such as
>
> dummy_backup_blocks = off (default) | on
>
> to explicitly indicate to the recovery process that backup blocks are
> present, yet they are garbage and should be ignored. Having garbage data
> within the system is potentially dangerous and I want to be told by the
> user that they were expecting that and its OK to ignore that data.
> Otherwise I want to throw informative errors. Maybe it seems OK now, but
> the next change to the system may have unintended consequences and it
> may not be us making the change. "It's OK the Alien will never escape
> from the lab" is the starting premise for many good sci-fi horrors and I
> want to watch them, not be in one myself. :-)
>
> We can call it other things, of course. e.g.
> ignore_dummy_blocks
> decompressed_blocks
> apply_backup_blocks

So far, we don't need any modification to the recovery and redo
functions.   They ignore the dummy and apply logical logs.   Also, if
there are both full page writes and logical log, current recovery
selects full page writes to apply.

I agree to introduce this option if 8.3 code introduces any conflict to
the current.   Or, we could introduce this option for future safety.  Do
you think we should introduce this option?

If this should be introduced now, what we should do is to check this
option when dummy full-page-write appears.


>
>> Yes I believe so. As pg_standby does not include any chance to meet
>> partial writes of pages, I believe you can omit all the full page
>> writes. Of course, as Tom Lange suggested in
>> http://archives.postgresql.org/pgsql-hackers/2007-02/msg00034.php
>> removing full page writes can lose a chance to recover from
>> partial/inconsisitent writes in the crash of pg_standby. In this case,
>> we have to import a backup and archive logs (with full page writes
>> during the backup) to recover. (We have to import them when the file
>> system crashes anyway). If it's okay, I believe
>> pg_compresslog/pg_decompresslog can be integrated with pg_standby.
>>
>> Maybe we can work together to include pg_compresslog/pg_decompresslog in
>> pg_standby.
>
> ISTM there are two options.
>
> I think this option is already possible:
>
> 1. Allow pg_decompresslog to operate on a file, replacing it with the
> expanded form, like gunzip, so we would do this:
>   restore_command = 'pg_standby %f decomp.tmp && pg_decompresslog
> decomp.tmp %p'
>
> though the decomp.tmp file would not get properly initialised or cleaned
> up when we finish.
>
> whereas this will take additional work
>
> 2. Allow pg_standby to write to stdin, so that we can do this:
>   restore_command = 'pg_standby %f | pg_decompresslog - %p'
>

Both seem to work fine.   pg_decompresslog will read entire file at the
beginning and so both two will be equivallent.   To make the second
option run quicker, pg_decompresslog needs some modification to read WAL
record one after another.

Anyway, could you try to run pg_standby with pg_compresslog and
pg_decompresslog?

----
Additional recomment on page header removal:

I found that it is not simple to keep page header in the compressed
archive log.   Because we eliminate unmarked full page writes and shift
the rest of the WAL file data, it is not simple to keep page header as
the page header in the compressed archive log.   It is much simpler to
remove page header as well and rebuild them.   I'd like to keep current
implementation in this point.

Any suggestions are welcome.
-------------------

I'll modify the name of the commands and post it hopefully within 20hours.

With Best Regards;

--
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Thu, 2007-03-29 at 17:50 +0900, Koichi Suzuki wrote:

> Not only full-page-writes are written as WAL record.   In my proposal,
> both full-page-writes and logical log are written in a WAL record, which
> will make WAL size slightly bigger (five percent or so).   If
> full_page_compress = off, only a full-page-write will be written in a
> WAL record.   I thought someone will not be happy with this size growth.

OK, I see what you're doing now and agree with you that we do need a
parameter. Not sure about the name you've chosen though - it certainly
confused me until you explained.

A parameter called ..._compress indicates to me that it would reduce
something in size whereas what it actually does is increase the size of
WAL slightly. We should have a parameter name that indicates what it
actually does, otherwise some people will choose to use this parameter
even when they are not using archive_command with pg_compresslog.

Some possible names...

additional_wal_info = 'COMPRESS'
add_wal_info
wal_additional_info
wal_auxiliary_info
wal_extra_data
attach_wal_info
...
others?

I've got some ideas for the future for adding additional WAL info for
various purposes, so it might be useful to have a parameter that can
cater for multiple types of additional WAL data. Or maybe we go for
something more specific like

wal_add_compress_info = on
wal_add_XXXX_info ...

> > In recovery.conf, I'd like to see a parameter such as
> >
> > dummy_backup_blocks = off (default) | on
> >
> > to explicitly indicate to the recovery process that backup blocks are
> > present, yet they are garbage and should be ignored. Having garbage data
> > within the system is potentially dangerous and I want to be told by the
> > user that they were expecting that and its OK to ignore that data.
> > Otherwise I want to throw informative errors. Maybe it seems OK now, but
> > the next change to the system may have unintended consequences and it
> > may not be us making the change. "It's OK the Alien will never escape
> > from the lab" is the starting premise for many good sci-fi horrors and I
> > want to watch them, not be in one myself. :-)
> >
> > We can call it other things, of course. e.g.
> > ignore_dummy_blocks
> > decompressed_blocks
> > apply_backup_blocks
>
> So far, we don't need any modification to the recovery and redo
> functions.   They ignore the dummy and apply logical logs.   Also, if
> there are both full page writes and logical log, current recovery
> selects full page writes to apply.
>
> I agree to introduce this option if 8.3 code introduces any conflict to
> the current.   Or, we could introduce this option for future safety.  Do
> you think we should introduce this option?

Yes. You are skipping a correctness test and that should be by explicit
command only. It's no problem to include that as well, since you are
already having to specify pg_... decompress... but the recovery process
doesn't know whether or not you've done that.

> Anyway, could you try to run pg_standby with pg_compresslog and
> pg_decompresslog?

After freeze, yes.

> ----
> Additional recomment on page header removal:
>
> I found that it is not simple to keep page header in the compressed
> archive log.   Because we eliminate unmarked full page writes and shift
> the rest of the WAL file data, it is not simple to keep page header as
> the page header in the compressed archive log.   It is much simpler to
> remove page header as well and rebuild them.   I'd like to keep current
> implementation in this point.

OK.

This is a good feature. Thanks for your patience with my comments.

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Josh Berkus
Date:
Simon,

> OK, different question:
> Why would anyone ever set full_page_compress = off?

The only reason I can see is if compression costs us CPU but gains RAM &
I/O.  I can think of a lot of applications ... benchmarks included ...
which are CPU-bound but not RAM or I/O bound.  For those applications,
compression is a bad tradeoff.

If, however, CPU used for compression is made up elsewhere through smaller
file processing, then I'd agree that we don't need a switch.

--
--Josh

Josh Berkus
PostgreSQL @ Sun
San Francisco

Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Thu, 2007-03-29 at 11:45 -0700, Josh Berkus wrote:

> > OK, different question:
> > Why would anyone ever set full_page_compress = off?
>
> The only reason I can see is if compression costs us CPU but gains RAM &
> I/O.  I can think of a lot of applications ... benchmarks included ...
> which are CPU-bound but not RAM or I/O bound.  For those applications,
> compression is a bad tradeoff.
>
> If, however, CPU used for compression is made up elsewhere through smaller
> file processing, then I'd agree that we don't need a switch.

Koichi-san has explained things for me now.

I misunderstood what the parameter did and reading your post, ISTM you
have as well. I do hope Koichi-san will alter the name to allow
everybody to understand what it does.

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Josh;

I'd like to explain what the term "compression" in my proposal means
again and would like to show the resource consumption comparision with
cp and gzip.

My proposal is to remove unnecessary full page writes (they are needed
in crash recovery from inconsistent or partial writes) when we copy WAL
to archive log and rebuilt them as a dummy when we restore from archive
log.  Dummy is needed to maintain LSN.  So it is very very different
from general purpose compression such as gzip, although pg_compresslog
compresses archive log as a result.

As to CPU and I/O consumption, I've already evaluated as follows:

1) Collect all the WAL segment.
2) Copy them by different means, cp, pg_compresslog and gzip.

and compared the ellapsed time as well as other resource consumption.

Benchmark: DBT-2
Database size: 120WH (12.3GB)
Total WAL size: 4.2GB (after 60min. run)
Elapsed time:
   cp:            120.6sec
   gzip:          590.0sec
   pg_compresslog: 79.4sec
Resultant archive log size:
   cp:             4.2GB
   gzip:           2.2GB
   pg_compresslog: 0.3GB
Resource consumption:
   cp:   user:   0.5sec system: 15.8sec idle:  16.9sec I/O wait: 87.7sec
   gzip: user: 286.2sec system:  8.6sec idle: 260.5sec I/O wait: 36.0sec
   pg_compresslog:
         user:   7.9sec system:  5.5sec idle:  37.8sec I/O wait: 28.4sec

Because the resultant log size is considerably smaller than cp or gzip,
pg_compresslog need much less I/O and because the logic is much simpler
than gzip, it does not consume CPU.

The term "compress" may not be appropriate.   We may call this "log
optimization" instead.

So I don't see any reason why this (at least optimization "mark" in each
log record) can't be integrated.

Simon Riggs wrote:
> On Thu, 2007-03-29 at 11:45 -0700, Josh Berkus wrote:
>
>>> OK, different question:
>>> Why would anyone ever set full_page_compress = off?
>> The only reason I can see is if compression costs us CPU but gains RAM &
>> I/O.  I can think of a lot of applications ... benchmarks included ...
>> which are CPU-bound but not RAM or I/O bound.  For those applications,
>> compression is a bad tradeoff.
>>
>> If, however, CPU used for compression is made up elsewhere through smaller
>> file processing, then I'd agree that we don't need a switch.

As I wrote to Simon's comment, I concern only one thing.

Without a switch, because both full page writes and corresponding
logical log is included in WAL, this will increase WAL size slightly
(maybe about five percent or so).   If everybody is happy with this, we
don't need a switch.

>
> Koichi-san has explained things for me now.
>
> I misunderstood what the parameter did and reading your post, ISTM you
> have as well. I do hope Koichi-san will alter the name to allow
> everybody to understand what it does.
>

Here're some candidates:
full_page_writes_optimize
full_page_writes_mark: means it marks full_page_write as "needed in
crash recovery", "needed in archive recovery" and so on.

I don't insist these names.  It's very helpful if you have any
suggestion to reflect what it really means.

Regards;
--
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

Here's a patch reflected some of Simon's comments.

1) Removed an elog call in a critical section.

2) Changed the name of the commands, pg_complesslog and pg_decompresslog.

3) Changed diff option to make a patch.

--
Koichi Suzuki

Attachment

Re: [PATCHES] Full page writes improvement, code update

From
"Zeugswetter Andreas ADI SD"
Date:
> Without a switch, because both full page writes and
> corresponding logical log is included in WAL, this will
> increase WAL size slightly
> (maybe about five percent or so).   If everybody is happy
> with this, we
> don't need a switch.

Sorry, I still don't understand that. What is the "corresponding logical
log" ?
It seems to me, that a full page WAL record has enough info to produce a

dummy LSN WAL entry. So insead of just cutting the full page wal record
you
could replace it with a LSN WAL entry when archiving the log.

Then all that is needed is the one flag, no extra space ?

Andreas


Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Fri, 2007-03-30 at 10:22 +0200, Zeugswetter Andreas ADI SD wrote:
> > Without a switch, because both full page writes and
> > corresponding logical log is included in WAL, this will
> > increase WAL size slightly
> > (maybe about five percent or so).   If everybody is happy
> > with this, we
> > don't need a switch.
>
> Sorry, I still don't understand that. What is the "corresponding logical
> log" ?
> It seems to me, that a full page WAL record has enough info to produce a
>
> dummy LSN WAL entry. So insead of just cutting the full page wal record
> you
> could replace it with a LSN WAL entry when archiving the log.
>
> Then all that is needed is the one flag, no extra space ?

The full page write is required for crash recovery, but that isn't
required during archive recovery because the base backup provides the
safe base. Archive recovery needs the normal xlog record, which in some
cases has been optimised away because the backup block is present, since
the full block already contains the changes.

If you want to remove the backup blocks, you need to put back the
information that was optimised away, otherwise you won't be able to do
the archive recovery correctly. Hence a slight increase in WAL volume to
allow it to be compressed does make sense.

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
"Zeugswetter Andreas ADI SD"
Date:
> Archive recovery needs the
> normal xlog record, which in some cases has been optimised
> away because the backup block is present, since the full
> block already contains the changes.

Aah, I didn't know that optimization exists.
I agree that removing that optimization is good/ok.

Andreas


Re: [PATCHES] Full page writes improvement, code update

From
Richard Huxton
Date:
Simon Riggs wrote:
> On Fri, 2007-03-30 at 10:22 +0200, Zeugswetter Andreas ADI SD wrote:
>>> Without a switch, because both full page writes and
>>> corresponding logical log is included in WAL, this will
>>> increase WAL size slightly
>>> (maybe about five percent or so).   If everybody is happy
>>> with this, we
>>> don't need a switch.
>> Sorry, I still don't understand that. What is the "corresponding logical
>> log" ?
>> It seems to me, that a full page WAL record has enough info to produce a
>>
>> dummy LSN WAL entry. So insead of just cutting the full page wal record
>> you
>> could replace it with a LSN WAL entry when archiving the log.
>>
>> Then all that is needed is the one flag, no extra space ?
>
> The full page write is required for crash recovery, but that isn't
> required during archive recovery because the base backup provides the
> safe base.

Is that always true? Could the backup not pick up a partially-written
page? Assuming it's being written to as the backup is in progress. (We
are talking about when disk blocks are smaller than PG blocks here, so
can't guarantee an atomic write for a PG block?)

--
   Richard Huxton
   Archonet Ltd

Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Fri, 2007-03-30 at 11:27 +0100, Richard Huxton wrote:

> Is that always true? Could the backup not pick up a partially-written
> page? Assuming it's being written to as the backup is in progress. (We
> are talking about when disk blocks are smaller than PG blocks here, so
> can't guarantee an atomic write for a PG block?)

Any page written during a backup has a backup block that would not be
removable by Koichi's tool, so yes, you'd still be safe.

i.e. between pg_start_backup() and pg_stop_backup() we always use full
page writes, even if you are running in full_page_writes=off mode.

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Richard Huxton
Date:
Simon Riggs wrote:
> On Fri, 2007-03-30 at 11:27 +0100, Richard Huxton wrote:
>
>> Is that always true? Could the backup not pick up a partially-written
>> page? Assuming it's being written to as the backup is in progress. (We
>> are talking about when disk blocks are smaller than PG blocks here, so
>> can't guarantee an atomic write for a PG block?)
>
> Any page written during a backup has a backup block that would not be
> removable by Koichi's tool, so yes, you'd still be safe.
>
> i.e. between pg_start_backup() and pg_stop_backup() we always use full
> page writes, even if you are running in full_page_writes=off mode.

Ah, that's OK then.

--
   Richard Huxton
   Archonet Ltd

Re: [PATCHES] Full page writes improvement, code update

From
Tom Lane
Date:
"Simon Riggs" <simon@2ndquadrant.com> writes:
> Any page written during a backup has a backup block that would not be
> removable by Koichi's tool, so yes, you'd still be safe.

How does it know not to do that?

            regards, tom lane

Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Fri, 2007-03-30 at 16:35 -0400, Tom Lane wrote:
> "Simon Riggs" <simon@2ndquadrant.com> writes:
> > Any page written during a backup has a backup block that would not be
> > removable by Koichi's tool, so yes, you'd still be safe.
>
> How does it know not to do that?

Not sure what you mean, but I'll take a stab...

I originally questioned Koichi-san's request for a full_page_compress
parameter, which is how it would tell whether/not. After explanation, I
accepted the need for a parameter, but I think we're looking for a new
name for it.

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Tom Lane wrote:
> "Simon Riggs" <simon@2ndquadrant.com> writes:
>> Any page written during a backup has a backup block that would not be
>> removable by Koichi's tool, so yes, you'd still be safe.
>
> How does it know not to do that?
>
>             regards, tom lane
>
> ---------------------------(end of broadcast)---------------------------
> TIP 5: don't forget to increase your free space map settings
>

XLogInsert( ) already has a logic to determine if inserting WAL record
is between pg_start_backup and pg_stop_backup.   Currently it is used
to remove full_page_writes when full_page_writes=off.   We can use
this to mark WAL records.   We have one bit not used in WAL record
header, the last bit of xl_info, where upper four bits are used to
indicate the resource manager and three of the rest are used to
indicate number of full page writes included in the record.

In my proposal, this unused bit is used to mark that full page
writes must not be removed at offline optimization by pg_compresslog.

Regards;

--
------
Koichi Suzuki

--
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
Bruce Momjian
Date:
Your patch has been added to the PostgreSQL unapplied patches list at:

    http://momjian.postgresql.org/cgi-bin/pgpatches

It will be applied as soon as one of the PostgreSQL committers reviews
and approves it.

---------------------------------------------------------------------------


Koichi Suzuki wrote:
> Hi,
>
> Here's a patch reflected some of Simon's comments.
>
> 1) Removed an elog call in a critical section.
>
> 2) Changed the name of the commands, pg_complesslog and pg_decompresslog.
>
> 3) Changed diff option to make a patch.
>
> --
> Koichi Suzuki

[ Attachment, skipping... ]

>
> ---------------------------(end of broadcast)---------------------------
> TIP 5: don't forget to increase your free space map settings

--
  Bruce Momjian  <bruce@momjian.us>          http://momjian.us
  EnterpriseDB                               http://www.enterprisedb.com

  + If your life is a hard drive, Christ can be your backup. +

Re: [PATCHES] Full page writes improvement, code update again.

From
Koichi Suzuki
Date:
Here's third revision of WAL archival optimization patch.   GUC
parameter name was changed to wal_add_optimization_info.

Regards;

--
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Bruce Momjian wrote:
> Your patch has been added to the PostgreSQL unapplied patches list at:
>
>     http://momjian.postgresql.org/cgi-bin/pgpatches

Thank you very much for including.   Attached is an update of the patch
according to Simon Riggs's comment about GUC name.

Regards;

--
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
Bruce Momjian
Date:
Your patch has been added to the PostgreSQL unapplied patches list at:

    http://momjian.postgresql.org/cgi-bin/pgpatches

It will be applied as soon as one of the PostgreSQL committers reviews
and approves it.

---------------------------------------------------------------------------


Koichi Suzuki wrote:
> Bruce Momjian wrote:
> > Your patch has been added to the PostgreSQL unapplied patches list at:
> >
> >     http://momjian.postgresql.org/cgi-bin/pgpatches
>
> Thank you very much for including.   Attached is an update of the patch
> according to Simon Riggs's comment about GUC name.
>
> Regards;
>
> --
> Koichi Suzuki

[ Attachment, skipping... ]

>
> ---------------------------(end of broadcast)---------------------------
> TIP 6: explain analyze is your friend

--
  Bruce Momjian  <bruce@momjian.us>          http://momjian.us
  EnterpriseDB                               http://www.enterprisedb.com

  + If your life is a hard drive, Christ can be your backup. +

Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Tue, 2007-04-03 at 19:45 +0900, Koichi Suzuki wrote:
> Bruce Momjian wrote:
> > Your patch has been added to the PostgreSQL unapplied patches list at:
> >
> >     http://momjian.postgresql.org/cgi-bin/pgpatches
>
> Thank you very much for including.   Attached is an update of the patch
> according to Simon Riggs's comment about GUC name.

The patch comes with its own "install kit", which is great to review
(many thanks), but hard to determine where you think code should go when
committed.

My guess based on your patch
- the patch gets applied to core :-)
- pg_compresslog *and* pg_decompresslog go to a contrib directory called
contrib/lesslog?

Can you please produce a combined patch that does all of the above, plus
edits the contrib Makefile to add all of those, as well as editing the
README so it doesn't mention the patch, just the contrib executables?

The patch looks correct to me now. I haven't tested it yet, but will be
doing so in the last week of April, which is when I'll be doing docs for
this and other stuff, since time is pressing now.

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

I agree to put the patch to core and the others (pg_compresslog and
pg_decompresslog) to contrib/lesslog.

I will make separate materials to go to core and contrib.

As for patches, we have tested against pgbench, DBT-2 and our
propriatery benchmarks and it looked to run correctly.

Regards;

Simon Riggs wrote:
> On Tue, 2007-04-03 at 19:45 +0900, Koichi Suzuki wrote:
>> Bruce Momjian wrote:
>>> Your patch has been added to the PostgreSQL unapplied patches list at:
>>>
>>>     http://momjian.postgresql.org/cgi-bin/pgpatches
>> Thank you very much for including.   Attached is an update of the patch
>> according to Simon Riggs's comment about GUC name.
>
> The patch comes with its own "install kit", which is great to review
> (many thanks), but hard to determine where you think code should go when
> committed.
>
> My guess based on your patch
> - the patch gets applied to core :-)
> - pg_compresslog *and* pg_decompresslog go to a contrib directory called
> contrib/lesslog?
>
> Can you please produce a combined patch that does all of the above, plus
> edits the contrib Makefile to add all of those, as well as editing the
> README so it doesn't mention the patch, just the contrib executables?
>
> The patch looks correct to me now. I haven't tested it yet, but will be
> doing so in the last week of April, which is when I'll be doing docs for
> this and other stuff, since time is pressing now.
>


--
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
Tom Lane
Date:
Koichi Suzuki <suzuki.koichi@oss.ntt.co.jp> writes:
> My proposal is to remove unnecessary full page writes (they are needed
> in crash recovery from inconsistent or partial writes) when we copy WAL
> to archive log and rebuilt them as a dummy when we restore from archive
> log.
> ...
> Benchmark: DBT-2
> Database size: 120WH (12.3GB)
> Total WAL size: 4.2GB (after 60min. run)
> Elapsed time:
>    cp:            120.6sec
>    gzip:          590.0sec
>    pg_compresslog: 79.4sec
> Resultant archive log size:
>    cp:             4.2GB
>    gzip:           2.2GB
>    pg_compresslog: 0.3GB
> Resource consumption:
>    cp:   user:   0.5sec system: 15.8sec idle:  16.9sec I/O wait: 87.7sec
>    gzip: user: 286.2sec system:  8.6sec idle: 260.5sec I/O wait: 36.0sec
>    pg_compresslog:
>          user:   7.9sec system:  5.5sec idle:  37.8sec I/O wait: 28.4sec

What checkpoint settings were used to make this comparison?  I'm
wondering whether much of the same benefit can't be bought at zero cost
by increasing the checkpoint interval, because that translates directly
to a reduction in the number of full-page images inserted into WAL.

Also, how much was the database run itself slowed down by the increased
volume of WAL (due to duplicated information)?  It seems rather
pointless to me to measure only the archiving effort without any
consideration for the impact on the database server proper.

            regards, tom lane

PS: there's something fishy about the gzip numbers ... why all the idle
time?

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

In the case below, we run DBT-2 benchmark for one hour to get the
measure.   Checkpoint occured three times (checkpoint interval was 20min).

For more information, when checkpoint interval is one hour, the amount
of the archived log size was as follows:
cp:             3.1GB
gzip:           1.5GB
pg_compresslog: 0.3GB

For both cases, database size was 12.7GB, relatively small.

As pointed out, if we don't run the checkpoint forever, the value for cp
will become close to that for pg_compresslog, but it is not practical.

The point here is, if we collect archive log with cp and the average
work load is a quarter of the full power, cp archiving will produce
about 0.8GB archive log per hour (for DBT-2 case, of course the size
depends on the nature of the transaction).   If we run the database
whole day, the amount of the archive log will be as large as database
itself.   After one week, archive log size gets seven times as large as
the database itself.   This is the point.   In production, such large
archive log will raise storage cost.   The purpose of the proposal is
not to improve the performance, but to decrease the size of archive log
to save necessary storage, preserving the same chance of recovery at the
crash recovery as full_page_writes=on.

Because of DBT-2 nature, it is not meaningful to compare the throuput
(databsae size determines the number of transactions to run).   Instead,
 I compared the throuput using pgbench.   These measures are: cp:
570tps, gzip:558tps, pg_compresslog: 574tps, negligible difference.

In terms of idle time for gzip and other command to archive WAL offline,
no difference in the environment was given other than the command to
archive.   My guess is because the user time is very large in gzip, it
has more chance for scheduler to give resource to other processes.   In
the case of cp, idle time is more than 30times longer than user time.
Pg_compresslog uses seven times longer idle time than user time.  On the
other hand, gzip uses less idle time than user time.   Considering the
total amount of user time, I think it's reasonable measure.

Again, in my proposal, it is not the issue to increase run time
performance.   Issue is to decrease the size of archive log to save the
storage.

Regards;


Tom Lane wrote:
> Koichi Suzuki <suzuki.koichi@oss.ntt.co.jp> writes:
>> My proposal is to remove unnecessary full page writes (they are needed
>> in crash recovery from inconsistent or partial writes) when we copy WAL
>> to archive log and rebuilt them as a dummy when we restore from archive
>> log.
>> ...
>> Benchmark: DBT-2
>> Database size: 120WH (12.3GB)
>> Total WAL size: 4.2GB (after 60min. run)
>> Elapsed time:
>>    cp:            120.6sec
>>    gzip:          590.0sec
>>    pg_compresslog: 79.4sec
>> Resultant archive log size:
>>    cp:             4.2GB
>>    gzip:           2.2GB
>>    pg_compresslog: 0.3GB
>> Resource consumption:
>>    cp:   user:   0.5sec system: 15.8sec idle:  16.9sec I/O wait: 87.7sec
>>    gzip: user: 286.2sec system:  8.6sec idle: 260.5sec I/O wait: 36.0sec
>>    pg_compresslog:
>>          user:   7.9sec system:  5.5sec idle:  37.8sec I/O wait: 28.4sec
>
> What checkpoint settings were used to make this comparison?  I'm
> wondering whether much of the same benefit can't be bought at zero cost
> by increasing the checkpoint interval, because that translates directly
> to a reduction in the number of full-page images inserted into WAL.
>
> Also, how much was the database run itself slowed down by the increased
> volume of WAL (due to duplicated information)?  It seems rather
> pointless to me to measure only the archiving effort without any
> consideration for the impact on the database server proper.
>
>             regards, tom lane
>
> PS: there's something fishy about the gzip numbers ... why all the idle
> time?
>
> ---------------------------(end of broadcast)---------------------------
> TIP 4: Have you searched our list archives?
>
>                http://archives.postgresql.org
>


--
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
"Joshua D. Drake"
Date:
> In terms of idle time for gzip and other command to archive WAL offline,
> no difference in the environment was given other than the command to
> archive.   My guess is because the user time is very large in gzip, it
> has more chance for scheduler to give resource to other processes.   In
> the case of cp, idle time is more than 30times longer than user time.
> Pg_compresslog uses seven times longer idle time than user time.  On the
> other hand, gzip uses less idle time than user time.   Considering the
> total amount of user time, I think it's reasonable measure.
>
> Again, in my proposal, it is not the issue to increase run time
> performance.   Issue is to decrease the size of archive log to save the
> storage.

Considering the relatively little amount of storage a transaction log
takes, it would seem to me that the performance angle is more appropriate.

Is it more efficient in other ways besides negligible tps? Possibly more
efficient memory usage? Better restore times for a crashed system?

Sincerely,

Joshua D. Drake

>
> Regards;
>


--

      === The PostgreSQL Company: Command Prompt, Inc. ===
Sales/Support: +1.503.667.4564 || 24x7/Emergency: +1.800.492.2240
Providing the most comprehensive  PostgreSQL solutions since 1997
             http://www.commandprompt.com/

Donate to the PostgreSQL Project: http://www.postgresql.org/about/donate
PostgreSQL Replication: http://www.commandprompt.com/products/


Re: [PATCHES] Full page writes improvement, code update

From
Hannu Krosing
Date:
Ühel kenal päeval, T, 2007-04-10 kell 18:17, kirjutas Joshua D. Drake:
> > In terms of idle time for gzip and other command to archive WAL offline,
> > no difference in the environment was given other than the command to
> > archive.   My guess is because the user time is very large in gzip, it
> > has more chance for scheduler to give resource to other processes.   In
> > the case of cp, idle time is more than 30times longer than user time.
> > Pg_compresslog uses seven times longer idle time than user time.  On the
> > other hand, gzip uses less idle time than user time.   Considering the
> > total amount of user time, I think it's reasonable measure.
> >
> > Again, in my proposal, it is not the issue to increase run time
> > performance.   Issue is to decrease the size of archive log to save the
> > storage.
>
> Considering the relatively little amount of storage a transaction log
> takes, it would seem to me that the performance angle is more appropriate.

As I understand it it's not about transaction log but about write-ahead
log.

and the amount of data in WAL can become very important once you have to
keep standby servers in different physical locations (cities, countries
or continents) where channel throughput and cost comes into play.

With simple cp (scp/rsync) the amount of WAL data needing to be copied
is about 10x more than data collected by trigger based solutions
(Slony/pgQ). With pg_compresslog WAL-shipping seems to have roughly the
same amount and thus becomes a viable alternative again.

> Is it more efficient in other ways besides negligible tps? Possibly more
> efficient memory usage? Better restore times for a crashed system?

I think that TPS is more affected by number of writes than size of each
block written, so there is probably not that much to gain in TPS, except
perhaps from better disk cache usage.

For me pg_compresslog seems to be a winner even if it just does not
degrade performance.

--
----------------
Hannu Krosing
Database Architect
Skype Technologies OÜ
Akadeemia tee 21 F, Tallinn, 12618, Estonia

Skype me:  callto:hkrosing
Get Skype for free:  http://www.skype.com



Re: [PATCHES] Full page writes improvement, code update

From
Tom Lane
Date:
Koichi Suzuki <suzuki.koichi@oss.ntt.co.jp> writes:
> For more information, when checkpoint interval is one hour, the amount
> of the archived log size was as follows:
> cp:             3.1GB
> gzip:           1.5GB
> pg_compresslog: 0.3GB

The notion that 90% of the WAL could be backup blocks even at very long
checkpoint intervals struck me as excessive, so I went looking for a
reason, and I may have found one.  There has been a bug in CVS HEAD
since Feb 8 causing every btree page split record to include a backup
block whether needed or not.  If these numbers were taken with recent
8.3 code, please retest with current HEAD.

            regards, tom lane

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
The score below was taken based on 8.2 code, not 8.3 code.  So I don't
think the below measure is introduced only in 8.3 code.

Tom Lane wrote:
> Koichi Suzuki <suzuki.koichi@oss.ntt.co.jp> writes:
>> For more information, when checkpoint interval is one hour, the amount
>> of the archived log size was as follows:
>> cp:             3.1GB
>> gzip:           1.5GB
>> pg_compresslog: 0.3GB
>
> The notion that 90% of the WAL could be backup blocks even at very long
> checkpoint intervals struck me as excessive, so I went looking for a
> reason, and I may have found one.  There has been a bug in CVS HEAD
> since Feb 8 causing every btree page split record to include a backup
> block whether needed or not.  If these numbers were taken with recent
> 8.3 code, please retest with current HEAD.
>
>             regards, tom lane
>


--
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
I don't fully understand what "transaction log" means.   If it means
"archived WAL", the current (8.2) code handle WAL as follows:

1) If full_page_writes=off, then no full page writes will be written to
WAL, except for those during onlie backup (between pg_start_backup and
pg_stop_backup).   The WAL size will be considerably small but it cannot
recover from partial/inconsistent write to the database files.  We have
to go back to the online backup and apply all the archive log.

2) If full_page_writes=on, then full page writes will be written at the
first update of a page after each checkpoint, plus full page writes at
1).   Because we have no means (in 8.2) to optimize the WAL so far, what
we can do is to copy WAL or gzip it at archive time.

If we'd like to keep good chance of recovery after the crash, 8.2
provides only the method 2), leaving archive log size considerably
large.  My proposal maintains the chance of crash recovery the same as
in the case of full_page_writes=on and reduces the size of archived log
as in the case of full_page_writes=off.

Regards;

Hannu Krosing wrote:
> Ühel kenal päeval, T, 2007-04-10 kell 18:17, kirjutas Joshua D. Drake:
>>> In terms of idle time for gzip and other command to archive WAL offline,
>>> no difference in the environment was given other than the command to
>>> archive.   My guess is because the user time is very large in gzip, it
>>> has more chance for scheduler to give resource to other processes.   In
>>> the case of cp, idle time is more than 30times longer than user time.
>>> Pg_compresslog uses seven times longer idle time than user time.  On the
>>> other hand, gzip uses less idle time than user time.   Considering the
>>> total amount of user time, I think it's reasonable measure.
>>>
>>> Again, in my proposal, it is not the issue to increase run time
>>> performance.   Issue is to decrease the size of archive log to save the
>>> storage.
>> Considering the relatively little amount of storage a transaction log
>> takes, it would seem to me that the performance angle is more appropriate.
>
> As I understand it it's not about transaction log but about write-ahead
> log.
>
> and the amount of data in WAL can become very important once you have to
> keep standby servers in different physical locations (cities, countries
> or continents) where channel throughput and cost comes into play.
>
> With simple cp (scp/rsync) the amount of WAL data needing to be copied
> is about 10x more than data collected by trigger based solutions
> (Slony/pgQ). With pg_compresslog WAL-shipping seems to have roughly the
> same amount and thus becomes a viable alternative again.
>
>> Is it more efficient in other ways besides negligible tps? Possibly more
>> efficient memory usage? Better restore times for a crashed system?
>
> I think that TPS is more affected by number of writes than size of each
> block written, so there is probably not that much to gain in TPS, except
> perhaps from better disk cache usage.
>
> For me pg_compresslog seems to be a winner even if it just does not
> degrade performance.
>


--
Koichi Suzuki


Re: [PATCHES] Full page writes improvement, code update

From
"Zeugswetter Andreas ADI SD"
Date:
> I don't fully understand what "transaction log" means.   If it means
> "archived WAL", the current (8.2) code handle WAL as follows:

Probably we can define "transaction log" to be the part of WAL that is
not
full pages.

> 1) If full_page_writes=off, then no full page writes will be
> written to WAL, except for those during onlie backup (between
> pg_start_backup and
> pg_stop_backup).   The WAL size will be considerably small
> but it cannot
> recover from partial/inconsistent write to the database
> files.  We have to go back to the online backup and apply all
> the archive log.
>
> 2) If full_page_writes=on, then full page writes will be
> written at the first update of a page after each checkpoint,
> plus full page writes at
> 1).   Because we have no means (in 8.2) to optimize the WAL
> so far, what
> we can do is to copy WAL or gzip it at archive time.
>
> If we'd like to keep good chance of recovery after the crash,
> 8.2 provides only the method 2), leaving archive log size
> considerably large.  My proposal maintains the chance of
> crash recovery the same as in the case of full_page_writes=on
> and reduces the size of archived log as in the case of
> full_page_writes=off.

Yup, this is a good summary.

You say you need to remove the optimization that avoids
the logging of a new tuple because the full page image exists.
I think we must already have the info in WAL which tuple inside the full
page image
is new (the one for which we avoided the WAL entry for).

How about this:
Leave current WAL as it is and only add the not removeable flag to full
pages.
pg_compresslog then replaces the full page image with a record for the
one tuple that is changed.
I tend to think it is not worth the increased complexity only to save
bytes in the uncompressed WAL though.

Another point about pg_decompresslog:

Why do you need a pg_decompresslog ? Imho pg_compresslog should already
do the replacing of the
full_page with the dummy entry. Then pg_decompresslog could be a simple
gunzip, or whatever compression was used,
but no logic.

Andreas

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

Sorry, inline reply.

Zeugswetter Andreas ADI SD wrote:

>
> Yup, this is a good summary.
>
> You say you need to remove the optimization that avoids
> the logging of a new tuple because the full page image exists.
> I think we must already have the info in WAL which tuple inside the full
> page image
> is new (the one for which we avoided the WAL entry for).
>
> How about this:
> Leave current WAL as it is and only add the not removeable flag to full
> pages.
> pg_compresslog then replaces the full page image with a record for the
> one tuple that is changed.
> I tend to think it is not worth the increased complexity only to save
> bytes in the uncompressed WAL though.

It is essentially what my patch proposes.  My patch includes flag to
full page writes which "can be" removed.

> Another point about pg_decompresslog:
>
> Why do you need a pg_decompresslog ? Imho pg_compresslog should already
> do the replacing of the
> full_page with the dummy entry. Then pg_decompresslog could be a simple
> gunzip, or whatever compression was used,
> but no logic.

Just removing full page writes does not work.   If we shift the rest of
the WAL, then LSN becomes inconsistent in compressed archive logs which
pg_compresslog produces.   For recovery, we have to restore LSN as the
original WAL.   Pg_decompresslog restores removed full page writes as a
dumm records so that recovery redo functions won't be confused.

Regards;

>
> Andreas
>


--
-------------
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
"Zeugswetter Andreas ADI SD"
Date:
> > Yup, this is a good summary.
> >
> > You say you need to remove the optimization that avoids the logging
of
> > a new tuple because the full page image exists.
> > I think we must already have the info in WAL which tuple inside the
> > full page image is new (the one for which we avoided the WAL entry
> > for).
> >
> > How about this:
> > Leave current WAL as it is and only add the not removeable flag to
> > full pages.
> > pg_compresslog then replaces the full page image with a record for
the
> > one tuple that is changed.
> > I tend to think it is not worth the increased complexity only to
save
> > bytes in the uncompressed WAL though.
>
> It is essentially what my patch proposes.  My patch includes
> flag to full page writes which "can be" removed.

Ok, a flag that marks full page images that can be removed is perfect.

But you also turn off the optimization that avoids writing regular
WAL records when the info is already contained in a full-page image
(increasing the
uncompressed size of WAL).
It was that part I questioned. As already stated, maybe I should not
have because
it would be too complex to reconstruct a regular WAL record from the
full-page image.
But that code would also be needed for WAL based partial replication, so
if it where too
complicated we would eventually want a switch to turn off the
optimization anyway
(at least for heap page changes).

> > Another point about pg_decompresslog:
> >
> > Why do you need a pg_decompresslog ? Imho pg_compresslog should
> > already do the replacing of the full_page with the dummy entry. Then

> > pg_decompresslog could be a simple gunzip, or whatever compression
was
> > used, but no logic.
>
> Just removing full page writes does not work.   If we shift the rest
of
> the WAL, then LSN becomes inconsistent in compressed archive logs
which
> pg_compresslog produces.   For recovery, we have to restore LSN as the

> original WAL.   Pg_decompresslog restores removed full page writes as
a
> dumm records so that recovery redo functions won't be confused.

Ah sorry, I needed some pgsql/src/backend/access/transam/README reading.

LSN is the physical position of records in WAL. Thus your dummy record
size is equal to what you cut out of the original record.
What about disconnecting WAL LSN from physical WAL record position
during replay ?
Add simple short WAL records in pg_compresslog like: advance LSN by 8192
bytes.

Andreas

Re: [PATCHES] Full page writes improvement, code update

From
Tom Lane
Date:
"Zeugswetter Andreas ADI SD" <ZeugswetterA@spardat.at> writes:
> But you also turn off the optimization that avoids writing regular
> WAL records when the info is already contained in a full-page image
> (increasing the uncompressed size of WAL).
> It was that part I questioned.

That's what bothers me about this patch, too.  It will be increasing
the cost of writing WAL (more data -> more CRC computation and more
I/O, not to mention more contention for the WAL locks) which translates
directly to a server slowdown.

The main arguments that I could see against Andreas' alternative are:

1. Some WAL record types are arranged in a way that actually would not
permit the reconstruction of the short form from the long form, because
they throw away too much data when the full-page image is substituted.
An example that's fresh in my mind is that the current format of the
btree page split WAL record discards newitemoff in that case, so you
couldn't identify the inserted item in the page image.  Now this is only
saving two bytes in what's usually going to be a darn large record
anyway, and it complicates the code to do it, so I wouldn't cry if we
changed btree split to include newitemoff always.  But there might be
some other cases where more data is involved.  In any case, someone
would have to look through every single WAL record type to determine
whether reconstruction is possible and fix it if not.

2. The compresslog utility would have to have specific knowledge about
every compressible WAL record type, to know how to convert it to the
short format.  That means an ongoing maintenance commitment there.
I don't think this is unacceptable, simply because we need only teach
it about a few common record types, not everything under the sun ---
anything it doesn't know how to fix, just leave alone, and if it's an
uncommon record type it really doesn't matter.  (I guess that means
that we don't really have to do #1 for every last record type, either.)

So I don't think either of these is a showstopper.  Doing it this way
would certainly make the patch more acceptable, since the argument that
it might hurt rather than help performance in some cases would go away.

> What about disconnecting WAL LSN from physical WAL record position
> during replay ?
> Add simple short WAL records in pg_compresslog like: advance LSN by 8192
> bytes.

I don't care for that, as it pretty much destroys some of the more
important sanity checks that xlog replay does.  The page boundaries
need to match the records contained in them.  So I think we do need
to have pg_decompresslog insert dummy WAL entries to fill up the
space saved by omitting full pages.

            regards, tom lane

Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Fri, 2007-04-13 at 10:36 -0400, Tom Lane wrote:
> "Zeugswetter Andreas ADI SD" <ZeugswetterA@spardat.at> writes:
> > But you also turn off the optimization that avoids writing regular
> > WAL records when the info is already contained in a full-page image
> > (increasing the uncompressed size of WAL).
> > It was that part I questioned.

I think its right to question it, certainly.

> That's what bothers me about this patch, too.  It will be increasing
> the cost of writing WAL (more data -> more CRC computation and more
> I/O, not to mention more contention for the WAL locks) which translates
> directly to a server slowdown.

I don't really understand this concern. Koichi-san has included a
parameter setting that would prevent any change at all in the way WAL is
written. If you don't want this slight increase in WAL, don't enable it.
If you do enable it, you'll also presumably be compressing the xlog too,
which works much better than gzip using less CPU. So overall it saves
more than it costs, ISTM, and nothing at all if you choose not to use
it.

> The main arguments that I could see against Andreas' alternative are:
>
> 1. Some WAL record types are arranged in a way that actually would not
> permit the reconstruction of the short form from the long form, because
> they throw away too much data when the full-page image is substituted.
> An example that's fresh in my mind is that the current format of the
> btree page split WAL record discards newitemoff in that case, so you
> couldn't identify the inserted item in the page image.  Now this is only
> saving two bytes in what's usually going to be a darn large record
> anyway, and it complicates the code to do it, so I wouldn't cry if we
> changed btree split to include newitemoff always.  But there might be
> some other cases where more data is involved.  In any case, someone
> would have to look through every single WAL record type to determine
> whether reconstruction is possible and fix it if not.
>
> 2. The compresslog utility would have to have specific knowledge about
> every compressible WAL record type, to know how to convert it to the
> short format.  That means an ongoing maintenance commitment there.
> I don't think this is unacceptable, simply because we need only teach
> it about a few common record types, not everything under the sun ---
> anything it doesn't know how to fix, just leave alone, and if it's an
> uncommon record type it really doesn't matter.  (I guess that means
> that we don't really have to do #1 for every last record type, either.)
>
> So I don't think either of these is a showstopper.  Doing it this way
> would certainly make the patch more acceptable, since the argument that
> it might hurt rather than help performance in some cases would go away.

Yeh, its additional code paths, but it sounds like Koichi-san and
colleagues are going to be trail blazing any bugs there and will be
around to fix any more that emerge.

> > What about disconnecting WAL LSN from physical WAL record position
> > during replay ?
> > Add simple short WAL records in pg_compresslog like: advance LSN by 8192
> > bytes.
>
> I don't care for that, as it pretty much destroys some of the more
> important sanity checks that xlog replay does.  The page boundaries
> need to match the records contained in them.  So I think we do need
> to have pg_decompresslog insert dummy WAL entries to fill up the
> space saved by omitting full pages.

Agreed. I don't want to start touching something that works so well.


We've been thinking about doing this for at least 3 years now, so I
don't see any reason to baulk at it now. I'm happy with Koichi-san's
patch as-is, assuming further extensive testing will be carried out on
it during beta.

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Tom Lane
Date:
"Simon Riggs" <simon@2ndquadrant.com> writes:
> On Fri, 2007-04-13 at 10:36 -0400, Tom Lane wrote:
>> That's what bothers me about this patch, too.  It will be increasing
>> the cost of writing WAL (more data -> more CRC computation and more
>> I/O, not to mention more contention for the WAL locks) which translates
>> directly to a server slowdown.

> I don't really understand this concern.

The real objection is that a patch that's alleged to make WAL smaller
actually does the exact opposite.  Now maybe you can buy that back
downstream of the archiver --- after yet more added-on processing ---
but it still seems that there's a fundamental misdesign here.

> Koichi-san has included a parameter setting that would prevent any
> change at all in the way WAL is written.

It bothers me that we'd need to have such a switch.  That's just another
way to shoot yourself in the foot, either by not enabling it (in which
case applying pg_compresslog as it stands would actively break your
WAL), or by enabling it when you weren't actually going to use
pg_compresslog (because you misunderstood the documentation to imply
that it'd make your WAL smaller by itself).  What I want to see is a
patch that doesn't bloat WAL at all and therefore doesn't need a switch.
I think Andreas is correct to complain that it should be done that way.

            regards, tom lane

Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Fri, 2007-04-13 at 11:47 -0400, Tom Lane wrote:
> "Simon Riggs" <simon@2ndquadrant.com> writes:
> > On Fri, 2007-04-13 at 10:36 -0400, Tom Lane wrote:
> >> That's what bothers me about this patch, too.  It will be increasing
> >> the cost of writing WAL (more data -> more CRC computation and more
> >> I/O, not to mention more contention for the WAL locks) which translates
> >> directly to a server slowdown.
>
> > I don't really understand this concern.
>
> The real objection is that a patch that's alleged to make WAL smaller
> actually does the exact opposite.  Now maybe you can buy that back
> downstream of the archiver --- after yet more added-on processing ---
> but it still seems that there's a fundamental misdesign here.
>
> > Koichi-san has included a parameter setting that would prevent any
> > change at all in the way WAL is written.
>
> It bothers me that we'd need to have such a switch.  That's just another
> way to shoot yourself in the foot, either by not enabling it (in which
> case applying pg_compresslog as it stands would actively break your
> WAL), or by enabling it when you weren't actually going to use
> pg_compresslog (because you misunderstood the documentation to imply
> that it'd make your WAL smaller by itself).  What I want to see is a
> patch that doesn't bloat WAL at all and therefore doesn't need a switch.
> I think Andreas is correct to complain that it should be done that way.

I agree with everything you say because we already had *exactly* this
discussion when the patch was already submitted, with me saying
everything you just said.

After a few things have been renamed to show their correct function and
impact, I am now comfortable with this patch.

Writing lots of additional code simply to remove a parameter that
*might* be mis-interpreted doesn't sound useful to me, especially when
bugs may leak in that way. My take is that this is simple and useful
*and* we have it now; other ways don't yet exist, nor will they in time
for 8.3.

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Tom Lane
Date:
"Simon Riggs" <simon@2ndquadrant.com> writes:
> Writing lots of additional code simply to remove a parameter that
> *might* be mis-interpreted doesn't sound useful to me, especially when
> bugs may leak in that way. My take is that this is simple and useful
> *and* we have it now; other ways don't yet exist, nor will they in time
> for 8.3.

The potential for misusing the switch is only one small part of the
argument; the larger part is that this has been done in the wrong way
and will cost performance unnecessarily.  The fact that it's ready
now is not something that I think should drive our choices.

I believe that it would be possible to make the needed core-server
changes in time for 8.3, and then to work on compress/decompress
on its own time scale and publish it on pgfoundry; with the hope
that it would be merged to contrib or core in 8.4.  Frankly the
compress/decompress code needs work anyway before it could be
merged (eg, I noted a distinct lack of I/O error checking).

            regards, tom lane

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Sorry I was very late to find this.

With DBT-2 benchmark, I've already compared the amount of WAL.   The
result was as follows:

Amount of WAL after 60min. run of DBT-2 benchmark
wal_add_optimization_info = off (default) 3.13GB
wal_add_optimization_info = on (new case) 3.17GB -> can be optimized to
0.31GB by pg_compresslog.

So the difference will be around a couple of percents.   I think this is
very good figure.

For information,
DB Size: 12.35GB (120WH)
Checkpoint timeout: 60min.  Checkpoint occured only once in the run.

----------------

I don't think replacing LSN works fine.  For full recovery to the
current time, we need both archive log and WAL.  Replacing LSN will make
archive log LSN inconsistent with WAL's LSN and the recovery will not work.

Reconstruction to regular WAL is proposed as pg_decompresslog.  We
should be careful enough not to make redo routines confused with the
dummy full page writes, as Simon suggested.  So far, it works fine.

Regards;

Zeugswetter Andreas ADI SD wrote:
>>> Yup, this is a good summary.
>>>
>>> You say you need to remove the optimization that avoids the logging
> of
>>> a new tuple because the full page image exists.
>>> I think we must already have the info in WAL which tuple inside the
>>> full page image is new (the one for which we avoided the WAL entry
>>> for).
>>>
>>> How about this:
>>> Leave current WAL as it is and only add the not removeable flag to
>>> full pages.
>>> pg_compresslog then replaces the full page image with a record for
> the
>>> one tuple that is changed.
>>> I tend to think it is not worth the increased complexity only to
> save
>>> bytes in the uncompressed WAL though.
>> It is essentially what my patch proposes.  My patch includes
>> flag to full page writes which "can be" removed.
>
> Ok, a flag that marks full page images that can be removed is perfect.
>
> But you also turn off the optimization that avoids writing regular
> WAL records when the info is already contained in a full-page image
> (increasing the
> uncompressed size of WAL).
> It was that part I questioned. As already stated, maybe I should not
> have because
> it would be too complex to reconstruct a regular WAL record from the
> full-page image.
> But that code would also be needed for WAL based partial replication, so
> if it where too
> complicated we would eventually want a switch to turn off the
> optimization anyway
> (at least for heap page changes).
>
>>> Another point about pg_decompresslog:
>>>
>>> Why do you need a pg_decompresslog ? Imho pg_compresslog should
>>> already do the replacing of the full_page with the dummy entry. Then
>
>>> pg_decompresslog could be a simple gunzip, or whatever compression
> was
>>> used, but no logic.
>> Just removing full page writes does not work.   If we shift the rest
> of
>> the WAL, then LSN becomes inconsistent in compressed archive logs
> which
>> pg_compresslog produces.   For recovery, we have to restore LSN as the
>
>> original WAL.   Pg_decompresslog restores removed full page writes as
> a
>> dumm records so that recovery redo functions won't be confused.
>
> Ah sorry, I needed some pgsql/src/backend/access/transam/README reading.
>
> LSN is the physical position of records in WAL. Thus your dummy record
> size is equal to what you cut out of the original record.
> What about disconnecting WAL LSN from physical WAL record position
> during replay ?
> Add simple short WAL records in pg_compresslog like: advance LSN by 8192
> bytes.
>
> Andreas
>


--
-------------
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
"Zeugswetter Andreas ADI SD"
Date:
> With DBT-2 benchmark, I've already compared the amount of WAL.   The
> result was as follows:
>
> Amount of WAL after 60min. run of DBT-2 benchmark
> wal_add_optimization_info = off (default) 3.13GB

how about wal_fullpage_optimization = on (default)

> wal_add_optimization_info = on (new case) 3.17GB -> can be
> optimized to 0.31GB by pg_compresslog.
>
> So the difference will be around a couple of percents.   I think this
is
> very good figure.
>
> For information,
> DB Size: 12.35GB (120WH)
> Checkpoint timeout: 60min.  Checkpoint occured only once in the run.

Unfortunately I think DBT-2 is not a good benchmark to test the disabled
wal optimization.
The test should contain some larger rows (maybe some updates on large
toasted values), and maybe more frequent checkpoints. Actually the poor
ratio between full pages and normal WAL content in this benchmark is
strange to begin with.
Tom fixed a bug recently, and it would be nice to see the new ratio.

Have you read Tom's comment on not really having to be able to
reconstruct all record types from the full page image ? I think that
sounded very promising (e.g. start out with only heap insert/update).

Then:
- we would not need the wal optimization switch (the full page flag
would always be added depending only on backup)
- pg_compresslog would only remove such "full page" images where it
knows how to reconstruct a "normal" WAL record from
- with time and effort pg_compresslog would be able to compress [nearly]
all record types's full images (no change in backend)

> I don't think replacing LSN works fine.  For full recovery to
> the current time, we need both archive log and WAL.
> Replacing LSN will make archive log LSN inconsistent with
> WAL's LSN and the recovery will not work.

WAL recovery would have had to be modified (decouple LSN from WAL
position during recovery).
An "archive log" would have been a valid WAL (with appropriate LSN
advance records).

> Reconstruction to regular WAL is proposed as
> pg_decompresslog.  We should be careful enough not to make
> redo routines confused with the dummy full page writes, as
> Simon suggested.  So far, it works fine.

Yes, Tom didn't like "LSN replacing" eighter. I withdraw my concern
regarding pg_decompresslog.

Your work in this area is extremely valuable and I hope my comments are
not discouraging.

Thank you
Andreas

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

I agree that pg_compresslog should be aware of all the WAL records'
details so that it can optimize archive log safely.   In my patch, I've
examined 8.2's WAL records to make pg_compresslog/pg_decompresslog safe.

Also I agree further pg_compresslog maintenance needs to examine changes
in WAL record format.   Because the number of such format will be
limited, I think the amount of work will be reasonable enough.

Regards;

Simon Riggs wrote:
> On Fri, 2007-04-13 at 10:36 -0400, Tom Lane wrote:
>> "Zeugswetter Andreas ADI SD" <ZeugswetterA@spardat.at> writes:
>>> But you also turn off the optimization that avoids writing regular
>>> WAL records when the info is already contained in a full-page image
>>> (increasing the uncompressed size of WAL).
>>> It was that part I questioned.
>
> I think its right to question it, certainly.
>
>> That's what bothers me about this patch, too.  It will be increasing
>> the cost of writing WAL (more data -> more CRC computation and more
>> I/O, not to mention more contention for the WAL locks) which translates
>> directly to a server slowdown.
>
> I don't really understand this concern. Koichi-san has included a
> parameter setting that would prevent any change at all in the way WAL is
> written. If you don't want this slight increase in WAL, don't enable it.
> If you do enable it, you'll also presumably be compressing the xlog too,
> which works much better than gzip using less CPU. So overall it saves
> more than it costs, ISTM, and nothing at all if you choose not to use
> it.
>
>> The main arguments that I could see against Andreas' alternative are:
>>
>> 1. Some WAL record types are arranged in a way that actually would not
>> permit the reconstruction of the short form from the long form, because
>> they throw away too much data when the full-page image is substituted.
>> An example that's fresh in my mind is that the current format of the
>> btree page split WAL record discards newitemoff in that case, so you
>> couldn't identify the inserted item in the page image.  Now this is only
>> saving two bytes in what's usually going to be a darn large record
>> anyway, and it complicates the code to do it, so I wouldn't cry if we
>> changed btree split to include newitemoff always.  But there might be
>> some other cases where more data is involved.  In any case, someone
>> would have to look through every single WAL record type to determine
>> whether reconstruction is possible and fix it if not.
>>
>> 2. The compresslog utility would have to have specific knowledge about
>> every compressible WAL record type, to know how to convert it to the
>> short format.  That means an ongoing maintenance commitment there.
>> I don't think this is unacceptable, simply because we need only teach
>> it about a few common record types, not everything under the sun ---
>> anything it doesn't know how to fix, just leave alone, and if it's an
>> uncommon record type it really doesn't matter.  (I guess that means
>> that we don't really have to do #1 for every last record type, either.)
>>
>> So I don't think either of these is a showstopper.  Doing it this way
>> would certainly make the patch more acceptable, since the argument that
>> it might hurt rather than help performance in some cases would go away.
>
> Yeh, its additional code paths, but it sounds like Koichi-san and
> colleagues are going to be trail blazing any bugs there and will be
> around to fix any more that emerge.
>
>>> What about disconnecting WAL LSN from physical WAL record position
>>> during replay ?
>>> Add simple short WAL records in pg_compresslog like: advance LSN by 8192
>>> bytes.
>> I don't care for that, as it pretty much destroys some of the more
>> important sanity checks that xlog replay does.  The page boundaries
>> need to match the records contained in them.  So I think we do need
>> to have pg_decompresslog insert dummy WAL entries to fill up the
>> space saved by omitting full pages.
>
> Agreed. I don't want to start touching something that works so well.
>
>
> We've been thinking about doing this for at least 3 years now, so I
> don't see any reason to baulk at it now. I'm happy with Koichi-san's
> patch as-is, assuming further extensive testing will be carried out on
> it during beta.
>


--
-------------
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Here's only a part of the reply I should do, but as to I/O error
checking ...

Here's a list of system calls and other external function/library calls
used in pg_lesslog patch series, together with how current patch checks
each errors and how current postgresql source handles the similar calls:

--------------------------------
1. No error check is done
1-1. fileno()
     fileno() is called against stdin and stdout from pg_compresslog.c
and pg_decompresslog.c.  They are intended to be invoked from a shell
and so stdin and stdout are both available.  fileno() error occurs only
if invoker of pg_compresslog or pg_decompresslog closes stdin and/or
stdout before the invoker executes them.   I found similar fileno()
usage in pg_dump/pg_backup_archive.c and postmaster/syslogger.c.   I
don't think this is an issue.

1-2. fflush()
    fflush() is called against stdout within a debug routine, debug.c.
Such usage can also be found in bin/initdb.c, bin/scripts/createdb.c,
bin/psql/common.c and more.   I don't think this is an issue either.

1-3. printf() and fprintf()
     It is common practice not to check the error.   We can find such
calls in many of existing source codes.

1-4. strerror()
     It is checked that system call returns error before calling
strerror.   Similar code can be found in other PostgreSQL source too.

----------------------------------
2. Error check is done
All the following function calls are associated with return value check.
open(), close(), fstat(), read(), write()

-----------------------------------
3. Functions do not return error
The following functin will not return errors, so no error check is needed.
exit(), memcopy(), memset(), strcmp()
------------------------------------

I hope this helps.

Regards;

Tom Lane wrote:
> "Simon Riggs" <simon@2ndquadrant.com> writes:
>> Writing lots of additional code simply to remove a parameter that
>> *might* be mis-interpreted doesn't sound useful to me, especially when
>> bugs may leak in that way. My take is that this is simple and useful
>> *and* we have it now; other ways don't yet exist, nor will they in time
>> for 8.3.
>
> The potential for misusing the switch is only one small part of the
> argument; the larger part is that this has been done in the wrong way
> and will cost performance unnecessarily.  The fact that it's ready
> now is not something that I think should drive our choices.
>
> I believe that it would be possible to make the needed core-server
> changes in time for 8.3, and then to work on compress/decompress
> on its own time scale and publish it on pgfoundry; with the hope
> that it would be merged to contrib or core in 8.4.  Frankly the
> compress/decompress code needs work anyway before it could be
> merged (eg, I noted a distinct lack of I/O error checking).
>
>             regards, tom lane
>


--
-------------
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
"Simon Riggs"
Date:
On Fri, 2007-04-20 at 10:16 +0200, Zeugswetter Andreas ADI SD wrote:

> Your work in this area is extremely valuable and I hope my comments are
> not discouraging.

I think its too late in the day to make the changes suggested by
yourself and Tom. They make the patch more invasive and more likely to
error, plus we don't have much time. This really means the patch is
likely to be rejected at the 11th hour when what we have essentially
works...

--
  Simon Riggs
  EnterpriseDB   http://www.enterprisedb.com



Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

I don't insist the name and the default of the GUC parameter.  I'm
afraid wal_fullpage_optimization = on (default) makes some confusion
because the default behavior becomes a bit different on WAL itself.

I'd like to have some more opinion on this.

Zeugswetter Andreas ADI SD wrote:
>> With DBT-2 benchmark, I've already compared the amount of WAL.   The
>> result was as follows:
>>
>> Amount of WAL after 60min. run of DBT-2 benchmark
>> wal_add_optimization_info = off (default) 3.13GB
>
> how about wal_fullpage_optimization = on (default)
>
>> wal_add_optimization_info = on (new case) 3.17GB -> can be
>> optimized to 0.31GB by pg_compresslog.
>>
>> So the difference will be around a couple of percents.   I think this
> is
>> very good figure.
>>
>> For information,
>> DB Size: 12.35GB (120WH)
>> Checkpoint timeout: 60min.  Checkpoint occured only once in the run.
>
> Unfortunately I think DBT-2 is not a good benchmark to test the disabled
> wal optimization.
> The test should contain some larger rows (maybe some updates on large
> toasted values), and maybe more frequent checkpoints. Actually the poor
> ratio between full pages and normal WAL content in this benchmark is
> strange to begin with.
> Tom fixed a bug recently, and it would be nice to see the new ratio.
>
> Have you read Tom's comment on not really having to be able to
> reconstruct all record types from the full page image ? I think that
> sounded very promising (e.g. start out with only heap insert/update).
>
> Then:
> - we would not need the wal optimization switch (the full page flag
> would always be added depending only on backup)
> - pg_compresslog would only remove such "full page" images where it
> knows how to reconstruct a "normal" WAL record from
> - with time and effort pg_compresslog would be able to compress [nearly]
> all record types's full images (no change in backend)
>
>> I don't think replacing LSN works fine.  For full recovery to
>> the current time, we need both archive log and WAL.
>> Replacing LSN will make archive log LSN inconsistent with
>> WAL's LSN and the recovery will not work.
>
> WAL recovery would have had to be modified (decouple LSN from WAL
> position during recovery).
> An "archive log" would have been a valid WAL (with appropriate LSN
> advance records).
>
>> Reconstruction to regular WAL is proposed as
>> pg_decompresslog.  We should be careful enough not to make
>> redo routines confused with the dummy full page writes, as
>> Simon suggested.  So far, it works fine.
>
> Yes, Tom didn't like "LSN replacing" eighter. I withdraw my concern
> regarding pg_decompresslog.
>
> Your work in this area is extremely valuable and I hope my comments are
> not discouraging.
>
> Thank you
> Andreas
>


--
-------------
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
"Zeugswetter Andreas ADI SD"
Date:
> I don't insist the name and the default of the GUC parameter.
>  I'm afraid wal_fullpage_optimization = on (default) makes
> some confusion because the default behavior becomes a bit
> different on WAL itself.

Seems my wal_fullpage_optimization is not a good name if it caused
misinterpretation already :-(

> >> Amount of WAL after 60min. run of DBT-2 benchmark
> >> wal_add_optimization_info = off (default) 3.13GB
> >
> > how about wal_fullpage_optimization = on (default)

The meaning of wal_fullpage_optimization = on (default)
would be the same as your wal_add_optimization_info = off (default).
(Reversed name, reversed meaning of the boolean value)

It would be there to *turn off* the (default) WAL full_page
optimization.
For your pg_compresslog it would need to be set to off.
"add_optimization_info" sounded like added info about/for some
optimization
which it is not. We turn off an optimization with the flag for the
benefit
of an easier pg_compresslog implementation.

As already said I would decouple this setting from the part that sets
the "removeable full page" flag in WAL, and making the recovery able to
skip dummy records. This I would do unconditionally.

Andreas

Re: [PATCHES] Full page writes improvement, code update

From
Josh Berkus
Date:
Hackers,

> Writing lots of additional code simply to remove a parameter that
> *might* be mis-interpreted doesn't sound useful to me, especially when
> bugs may leak in that way. My take is that this is simple and useful
> *and* we have it now; other ways don't yet exist, nor will they in time
> for 8.3.

How about naming the parameter wal_compressable?  That would indicate pretty
clearly that the parameter is intended to be used with wal_compress and
nothing else.

However, I do agree with Andreas that anything which adds to WAL volume, even
3%, seems like going in the wrong direction.  We already have higher log
output than any comparable database (higher than InnoDB by 3x) and we should
be looking for output to trim as well as compression.

So the relevant question is whether the patch in its current form provides
enough benefit to make it worthwhile for 8.3, or whether we should wait for
8.4.  Questions:

1) is there any throughput benefit for platforms with fast CPU but contrained
I/O (e.g. 2-drive webservers)?  Any penalty for servers with plentiful I/O?

2) Will this patch make attempts to reduce WAL volume in the future
significantly harder?

3) How is this better than command-line compression for log-shipping?  e.g.
why do we need it in the database?

--
Josh Berkus
PostgreSQL @ Sun
San Francisco

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

Sorry, because of so many comments/questions, I'll write inline....

Josh Berkus wrote:
> Hackers,
>
>> Writing lots of additional code simply to remove a parameter that
>> *might* be mis-interpreted doesn't sound useful to me, especially when
>> bugs may leak in that way. My take is that this is simple and useful
>> *and* we have it now; other ways don't yet exist, nor will they in time
>> for 8.3.
>
> How about naming the parameter wal_compressable?  That would indicate pretty
> clearly that the parameter is intended to be used with wal_compress and
> nothing else.

Hmm, it sounds nicer.

>
> However, I do agree with Andreas that anything which adds to WAL volume, even
> 3%, seems like going in the wrong direction.  We already have higher log
> output than any comparable database (higher than InnoDB by 3x) and we should
> be looking for output to trim as well as compression.
>
> So the relevant question is whether the patch in its current form provides
> enough benefit to make it worthwhile for 8.3, or whether we should wait for
> 8.4.  Questions:
>

Before answering questions below, I'd like to say that archive log
optimization has to be address different point of views to the current
(upto 8.2) settings.

1) To deal with partial/inconsisitent write to the data file at crash
recovery, we need full page writes at the first modification to pages
after each checkpoint.   It consumes much of WAL space.

2) 1) is not necessary for archive recovery (PITR) and full page writes
can be removed for this purpose.   However, we need full page writes
during hot backup to deal with partial writes by backup commands.  This
is implemented in 8.2.

3) To maintain crash recovery chance and reduce the amount of archive
log, removal of  unnecessary full page writes from archive logs is a
good choice.   To do this, we need both logical log and full page writes
in WAL.

I don't think there should be only one setting.   It depend on how
database is operated.   Leaving wal_add_optiomization_info = off default
does not bring any change in WAL and archive log handling.   I
understand some people may not be happy with additional 3% or so
increase in WAL size, especially people who dosn't need archive log at
all.   So I prefer to leave the default off.

For users, I think this is simple enough:

1) For people happy with 8.2 settings:
    No change is needed to move to 8.3 and there's really no change.

2) For people who need to reduce archive log size but like to leave full
page writes to WAL (to maintain crash recovery chance):
    a) Add GUC parameter: wal_add_optiomization_info=on
    b) Change archive command from "cp" to "pg_compresslog"
    c) Change restore command from "cp" to "pg_decompresslog"

Archive log can be stored and restored as done in older releases.

> 1) is there any throughput benefit for platforms with fast CPU but contrained
> I/O (e.g. 2-drive webservers)?  Any penalty for servers with plentiful I/O?

I've only run benchmarks with archive process running, because
wal_add_optimization_info=on does not make sense if we don't archive
WAL.   In this situation, total I/O decreases because writes to archive
log decreases.   Because of 3% or so increase in WAL size, there will be
increase in WAL write, but decrease in archive writes makes it up.

>
> 2) Will this patch make attempts to reduce WAL volume in the future
> significantly harder?

Yes, I'd like to continue to work to reduce the WAL size.   It's still
an issue when database size becomes several handreds of gigabytes in
size.   Anyway, I think WAL size reduction has to be done in
XLogInsert() or XLogWrite().   We need much more discussion for this.
The issue will be how to maintain crash recovery chance by inconsistent
writes (by full_page_writes=off, we have to give it up).   On the other
hand we have to keep examining each WAL record.

>
> 3) How is this better than command-line compression for log-shipping?  e.g.
> why do we need it in the database?

I don't fully understand what command-line compression means.   Simon
suggested that this patch can be used with log-shipping and I agree.
If we compare compression with gzip or other general purpose
compression, compression ratio, CPU usage and I/O by pg_compresslog are
all quite better than those in gzip.

Please let me know if you intended defferently.

Regards;

--
-------------
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
"Zeugswetter Andreas ADI SD"
Date:
> 3) To maintain crash recovery chance and reduce the amount of
> archive log, removal of  unnecessary full page writes from
> archive logs is a good choice.

Definitely, yes. pg_compresslog could even move the full pages written
during backup out of WAL and put them in a different file that needs to
be applied before replay of the corresponding WAL after a physical
restore. This would further help reduce log shipping volume.

> To do this, we need both logical log and full page writes in WAL.

This is only true in the sense, that it allows a less complex
implementation of pg_compresslog.

Basically a WAL record consists of info about what happened and
currently eighter per tuple new data or a full page image. The info of
"what happened" together with the full page image is sufficient to
reconstruct the "per tuple new data". There might be a few WAL record
types (e.g. in btree split ?) where this is not so, but we could eighter
fix those or not compress those.

This is why I don't like Josh's suggested name of wal_compressable
eighter.
WAL is compressable eighter way, only pg_compresslog would need to be
more complex if you don't turn off the full page optimization. I think a
good name would tell that you are turning off an optimization.
(thus my wal_fullpage_optimization on/off)

Andreas


Re: [PATCHES] Full page writes improvement, code update

From
Josh Berkus
Date:
Koichi, Andreas,

> 1) To deal with partial/inconsisitent write to the data file at crash
> recovery, we need full page writes at the first modification to pages
> after each checkpoint.   It consumes much of WAL space.

We need to find a way around this someday.  Other DBs don't do this; it may be
becuase they're less durable, or because they fixed the problem.

> I don't think there should be only one setting.   It depend on how
> database is operated.   Leaving wal_add_optiomization_info = off default
> does not bring any change in WAL and archive log handling.   I
> understand some people may not be happy with additional 3% or so
> increase in WAL size, especially people who dosn't need archive log at
> all.   So I prefer to leave the default off.

Except that, is there any reason to turn this off if we are archiving?  Maybe
it should just be slaved to archive_command ... if we're not using PITR, it's
off, if we are, it's on.

> > 1) is there any throughput benefit for platforms with fast CPU but
> > contrained I/O (e.g. 2-drive webservers)?  Any penalty for servers with
> > plentiful I/O?
>
> I've only run benchmarks with archive process running, because
> wal_add_optimization_info=on does not make sense if we don't archive
> WAL.   In this situation, total I/O decreases because writes to archive
> log decreases.   Because of 3% or so increase in WAL size, there will be
> increase in WAL write, but decrease in archive writes makes it up.

Yeah, I was just looking for a way to make this a performance feature.  I see
now that it can't be.  ;-)

> > 3) How is this better than command-line compression for log-shipping?
> > e.g. why do we need it in the database?
>
> I don't fully understand what command-line compression means.   Simon
> suggested that this patch can be used with log-shipping and I agree.
> If we compare compression with gzip or other general purpose
> compression, compression ratio, CPU usage and I/O by pg_compresslog are
> all quite better than those in gzip.

OK, that answered my question.

> This is why I don't like Josh's suggested name of wal_compressable
> eighter.
> WAL is compressable eighter way, only pg_compresslog would need to be
> more complex if you don't turn off the full page optimization. I think a
> good name would tell that you are turning off an optimization.
> (thus my wal_fullpage_optimization on/off)

Well, as a PG hacker I find the name wal_fullpage_optimization quite baffling
and I think our general user base will find it even more so.  Now that I have
Koichi's explanation of the problem, I vote for simply slaving this to the
PITR settings and not having a separate option at all.

--
Josh Berkus
PostgreSQL @ Sun
San Francisco

Re: [PATCHES] Full page writes improvement, code update

From
Tom Lane
Date:
Josh Berkus <josh@agliodbs.com> writes:
> Well, as a PG hacker I find the name wal_fullpage_optimization quite
> baffling and I think our general user base will find it even more so.
> Now that I have Koichi's explanation of the problem, I vote for simply
> slaving this to the PITR settings and not having a separate option at
> all.

The way to not have a separate option is to not need one, by having the
feature not cost anything extra in the first place.  Andreas and I have
made the point repeatedly about how to do that.

            regards, tom lane

Re: [PATCHES] Full page writes improvement, code update

From
"Zeugswetter Andreas ADI SD"
Date:
> > 1) To deal with partial/inconsisitent write to the data file at
crash
> > recovery, we need full page writes at the first modification to
pages
> > after each checkpoint.   It consumes much of WAL space.
>
> We need to find a way around this someday.  Other DBs don't
> do this; it may be becuase they're less durable, or because
> they fixed the problem.

They eighter can only detect a failure later (this may be a very long
time depending on access and verify runs) or they also write page
images. Those that write page images usually write "before images" to a
different area that is cleared periodically (e.g. during checkpoint).

Writing to a different area was considered in pg, but there were more
negative issues than positive.
So imho pg_compresslog is the correct path forward. The current
discussion is only about whether we want a more complex pg_compresslog
and no change to current WAL, or an increased WAL size for a less
complex implementation.
Both would be able to compress the WAL to the same "archive log" size.

Andreas

Re: [PATCHES] Full page writes improvement, code update

From
Kenneth Marshall
Date:
On Wed, Apr 25, 2007 at 10:00:16AM +0200, Zeugswetter Andreas ADI SD wrote:
>
> > > 1) To deal with partial/inconsisitent write to the data file at
> crash
> > > recovery, we need full page writes at the first modification to
> pages
> > > after each checkpoint.   It consumes much of WAL space.
> >
> > We need to find a way around this someday.  Other DBs don't
> > do this; it may be becuase they're less durable, or because
> > they fixed the problem.
>
> They eighter can only detect a failure later (this may be a very long
> time depending on access and verify runs) or they also write page
> images. Those that write page images usually write "before images" to a
> different area that is cleared periodically (e.g. during checkpoint).
>
> Writing to a different area was considered in pg, but there were more
> negative issues than positive.
> So imho pg_compresslog is the correct path forward. The current
> discussion is only about whether we want a more complex pg_compresslog
> and no change to current WAL, or an increased WAL size for a less
> complex implementation.
> Both would be able to compress the WAL to the same "archive log" size.
>
> Andreas
>
I definitely am in the camp of not increasing WAL size at all. If we
need a bit more complexity to ensure that, so be it. Any approach that
increases WAL volume would need to have an amazing benefit to make it
warranted. This certainly does not meet that criteria.

Ken


Re: [PATCHES] Full page writes improvement, code update

From
Josh Berkus
Date:
Andreas,

> Writing to a different area was considered in pg, but there were more
> negative issues than positive.
> So imho pg_compresslog is the correct path forward. The current
> discussion is only about whether we want a more complex pg_compresslog
> and no change to current WAL, or an increased WAL size for a less
> complex implementation.
> Both would be able to compress the WAL to the same "archive log" size.

Huh?  As conceived, pg_compresslog does nothing to lower log volume for
general purposes, just on-disk storage size for archiving.  It doesn't help
us at all with the tremendous amount of log we put out for an OLTP server,
for example.

Not that pg_compresslog isn't useful on its own for improving warm standby
managability, but it's completely separate from addressing the "we're logging
too much" issue.

--
Josh Berkus
PostgreSQL @ Sun
San Francisco

Re: [PATCHES] Full page writes improvement, code update

From
Tom Lane
Date:
Josh Berkus <josh@agliodbs.com> writes:
> Andreas,
>> So imho pg_compresslog is the correct path forward. The current
>> discussion is only about whether we want a more complex pg_compresslog
>> and no change to current WAL, or an increased WAL size for a less
>> complex implementation.
>> Both would be able to compress the WAL to the same "archive log" size.

> Huh?  As conceived, pg_compresslog does nothing to lower log volume for
> general purposes, just on-disk storage size for archiving.  It doesn't help
> us at all with the tremendous amount of log we put out for an OLTP server,
> for example.

I don't see how what you said refutes what he said.  The sticking point
here is that the patch as-proposed *increases* the log volume before
compression.

            regards, tom lane

Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

Zeugswetter Andreas ADI SD wrote:
>> I don't insist the name and the default of the GUC parameter.
>>  I'm afraid wal_fullpage_optimization = on (default) makes
>> some confusion because the default behavior becomes a bit
>> different on WAL itself.
>
> Seems my wal_fullpage_optimization is not a good name if it caused
> misinterpretation already :-(
>
>>>> Amount of WAL after 60min. run of DBT-2 benchmark
>>>> wal_add_optimization_info = off (default) 3.13GB
>>> how about wal_fullpage_optimization = on (default)
>
> The meaning of wal_fullpage_optimization = on (default)
> would be the same as your wal_add_optimization_info = off (default).
> (Reversed name, reversed meaning of the boolean value)
>
> It would be there to *turn off* the (default) WAL full_page
> optimization.
> For your pg_compresslog it would need to be set to off.
> "add_optimization_info" sounded like added info about/for some
> optimization
> which it is not. We turn off an optimization with the flag for the
> benefit
> of an easier pg_compresslog implementation.

For pg_compresslog to remove full page writes, we need
wal_add_optimization_info=on.

>
> As already said I would decouple this setting from the part that sets
> the "removeable full page" flag in WAL, and making the recovery able to
> skip dummy records. This I would do unconditionally.
>
> Andreas
>
> ---------------------------(end of broadcast)---------------------------
> TIP 3: Have you checked our extensive FAQ?
>
>                http://www.postgresql.org/docs/faq
>


--
-------------
Koichi Suzuki

Re: too much WAL volume

From
"Zeugswetter Andreas ADI SD"
Date:
> > Writing to a different area was considered in pg, but there were
more
> > negative issues than positive.
> > So imho pg_compresslog is the correct path forward. The current
> > discussion is only about whether we want a more complex
pg_compresslog
> > and no change to current WAL, or an increased WAL size for a less
> > complex implementation.
> > Both would be able to compress the WAL to the same "archive log"
size.
>
> Huh?  As conceived, pg_compresslog does nothing to lower log
> volume for general purposes, just on-disk storage size for
> archiving.  It doesn't help us at all with the tremendous
> amount of log we put out for an OLTP server, for example.

Ok, that is not related to the original discussion though.
I have thus changed the subject, and removed [PATCHES].

You cannot directly compare the pg WAL size with other db's since they
write parts to other areas (e.g. physical log in Informix). You would
need to include those writes in a fair comparison.
It is definitely not true, that writing to a different area has only
advantages. The consensus was, that writing the page images to the WAL
has more pro's. We could revisit the pros and cons though.

Other options involve special OS and hardware (we already have that), or
accepting a high risc of needing a
restore after power outage (we don't have that, because we use no
mechanism to detect such a failure).

I am not sure that shrinking per WAL record size (other than the full
page images), e.g. by only logging changed bytes and not whole tuples,
would have a huge impact on OLTP tx/sec, since the limiting factor is
IO's per second and not Mb per second. Recent developments like HOT seem
a lot more promising in this regard since they avoid IO.

Andreas


Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Josh,

Josh Berkus wrote:
> Koichi, Andreas,
>
>> 1) To deal with partial/inconsisitent write to the data file at crash
>> recovery, we need full page writes at the first modification to pages
>> after each checkpoint.   It consumes much of WAL space.
>
> We need to find a way around this someday.  Other DBs don't do this; it may be
> becuase they're less durable, or because they fixed the problem.

Maybe both.   Fixing the problem may need some means to detect
partial/inconsistent writes to the data files, which may needs
additional CPU resource.

>
>> I don't think there should be only one setting.   It depend on how
>> database is operated.   Leaving wal_add_optiomization_info = off default
>> does not bring any change in WAL and archive log handling.   I
>> understand some people may not be happy with additional 3% or so
>> increase in WAL size, especially people who dosn't need archive log at
>> all.   So I prefer to leave the default off.
>
> Except that, is there any reason to turn this off if we are archiving?  Maybe
> it should just be slaved to archive_command ... if we're not using PITR, it's
> off, if we are, it's on.

Hmm, this sounds to work.  On the other hand, existing users, who are
happy with the current archiving and would not like to change current
archiving command to pg_compresslog or archive log size will increase a
bit.  I'd like to hear some more on this.

>
>>> 1) is there any throughput benefit for platforms with fast CPU but
>>> contrained I/O (e.g. 2-drive webservers)?  Any penalty for servers with
>>> plentiful I/O?
>> I've only run benchmarks with archive process running, because
>> wal_add_optimization_info=on does not make sense if we don't archive
>> WAL.   In this situation, total I/O decreases because writes to archive
>> log decreases.   Because of 3% or so increase in WAL size, there will be
>> increase in WAL write, but decrease in archive writes makes it up.
>
> Yeah, I was just looking for a way to make this a performance feature.  I see
> now that it can't be.  ;-)

As to the performance feature, I tested the patch against 8.3HEAD.
With pgbench, throughput was as follows:
Case1. Archiver: cp command, wal_add_optimization_info = off,
        full_page_writes=on
Case2. Archiver: pg_compresslog, wal_add_optimization_info = on,
        full_page_writes=on
DB Size: 1.65GB, Total transaction:1,000,000

Throughput was:
Case1: 632.69TPS
Case2: 653.10TPS ... 3% gain.

Archive Log Size:
Case1: 1.92GB
Case2: 0.57GB (about 30% of the Case1)... Before compression, the size
was 1.92GB.  Because this is based on the number of WAL segment file
size, there will be at most 16MB error in the measurement.  If we count
this, the increase in WAL I/O will be less than 1%.


>
>>> 3) How is this better than command-line compression for log-shipping?
>>> e.g. why do we need it in the database?
>> I don't fully understand what command-line compression means.   Simon
>> suggested that this patch can be used with log-shipping and I agree.
>> If we compare compression with gzip or other general purpose
>> compression, compression ratio, CPU usage and I/O by pg_compresslog are
>> all quite better than those in gzip.
>
> OK, that answered my question.
>
>> This is why I don't like Josh's suggested name of wal_compressable
>> eighter.
>> WAL is compressable eighter way, only pg_compresslog would need to be
>> more complex if you don't turn off the full page optimization. I think a
>> good name would tell that you are turning off an optimization.
>> (thus my wal_fullpage_optimization on/off)
>
> Well, as a PG hacker I find the name wal_fullpage_optimization quite baffling
> and I think our general user base will find it even more so.  Now that I have
> Koichi's explanation of the problem, I vote for simply slaving this to the
> PITR settings and not having a separate option at all.

Could I have more specific suggestion on this?

Regards;


--
-------------
Koichi Suzuki

Re: too much WAL volume

From
Greg Smith
Date:
On Thu, 26 Apr 2007, Zeugswetter Andreas ADI SD wrote:

> I am not sure that shrinking per WAL record size (other than the full
> page images), e.g. by only logging changed bytes and not whole tuples,
> would have a huge impact on OLTP tx/sec, since the limiting factor is
> IO's per second and not Mb per second.

With the kind of caching controller that's necessary for any serious OLTP 
work with Postgres, number of I/Os per second isn't really an important 
number.  Total volume of writes to the WAL volume can be though.  It's 
difficult but not impossible to encounter a workload that becomes 
bottlenecked by WAL volume on a good OLTP server, particularly because 
that's often going to a single or RAID-1 disk.  Whether those workloads 
also have the appropriate properties such that their WAL could be shrunk 
usefully in real-time is a good question.

--
* Greg Smith gsmith@gregsmith.com http://www.gregsmith.com Baltimore, MD


Re: too much WAL volume

From
Jim Nasby
Date:
On Apr 27, 2007, at 4:58 AM, Greg Smith wrote:
> On Thu, 26 Apr 2007, Zeugswetter Andreas ADI SD wrote:
>> I am not sure that shrinking per WAL record size (other than the full
>> page images), e.g. by only logging changed bytes and not whole  
>> tuples,
>> would have a huge impact on OLTP tx/sec, since the limiting factor is
>> IO's per second and not Mb per second.
>
> With the kind of caching controller that's necessary for any  
> serious OLTP work with Postgres, number of I/Os per second isn't  
> really an important number.  Total volume of writes to the WAL  
> volume can be though.  It's difficult but not impossible to  
> encounter a workload that becomes bottlenecked by WAL volume on a  
> good OLTP server, particularly because that's often going to a  
> single or RAID-1 disk.  Whether those workloads also have the  
> appropriate properties such that their WAL could be shrunk usefully  
> in real-time is a good question.

Yes, but how many data drives would you need to have to bottleneck on  
WAL? Even if the entire database is memory resident you'd still have  
to write all the pages out at some point, and it seems to me that  
you'd need a fair amount of disk capacity the data directory before  
you got pegged by WAL.

When I did some DBT2 testing a bit over a year ago I had a 20 drive  
RAID10 for data and a mirror for WAL and was nowhere close to pegged  
on WAL (this was on a Sun V40 connected to one of their storage arrays).
--
Jim Nasby                                            jim@nasby.net
EnterpriseDB      http://enterprisedb.com      512.569.9461 (cell)




Re: too much WAL volume

From
Greg Smith
Date:
On Fri, 27 Apr 2007, Jim Nasby wrote:

> Yes, but how many data drives would you need to have to bottleneck on WAL? 
> Even if the entire database is memory resident you'd still have to write all 
> the pages out at some point, and it seems to me that you'd need a fair amount 
> of disk capacity the data directory before you got pegged by WAL.

Depends on the type of transactions.  If you're doing something with lots 
of INSERT and perhaps some UPDATE volume that doesn't need to read heavily 
from the database to complete, because most of the important stuff is 
already in memory, you might run into the WAL limit without too much on 
the database disk side.  I did say it was difficult...

> When I did some DBT2 testing a bit over a year ago I had a 20 drive RAID10 
> for data and a mirror for WAL and was nowhere close to pegged on WAL (this 
> was on a Sun V40 connected to one of their storage arrays).

No doubt, the main reason I haven't used DBT2 more is because the WAL 
volume produced before you run into database limited bottlenecks isn't 
large, and certainly not in the same ratio as some of the apps I'm 
modeling.  Mine lean more toward straightforward transaction logging in 
parts.

I'm running on similar hardware (V40 is very close, I think the EMC array 
I test against is a bit better than the most of the Sun models) and I've 
seen some scenarios that produce 40MB/s average - 60MB/s peak of WAL 
volume.  Sure seems like I'm rate limited by the RAID-1 WAL disk.  As you 
say, eventually all the data has to make it to disk, but since it's not 
too expensive nowadays to have gigabytes worth of memory and disk array 
cache you can put off database writes for a surprisingly long period of 
time with the right system design.  It's harder to buffer those pesky 
O_DIRECT WAL writes when they blow right though at least one level of 
cache.

--
* Greg Smith gsmith@gregsmith.com http://www.gregsmith.com Baltimore, MD


Re: [PATCHES] Full page writes improvement, code update

From
Koichi Suzuki
Date:
Hi,

As replied to "Patch queue triage" by Tom, here's simplified patch to
mark WAL record as "compressable", with no increase in WAL itself.
Compression/decompression commands will be posted separately to PG
Foundary for further review.

-------------------------------
As suggested by Tom, I agree that WAL should not include "both" full
page write and incremental (logical) log.   I began to examine WAL
record format to see if incremental log can be made from full page
writes.   It will be okay even before 8.4, if simplified patch to the
core is accepted.   I will post simplified patch to the core as follows:

1. Mark the flag to indicate that the WAL record is compressable from
full page writes to incremental log.  This flag will be set if
a) It is not written during the hot backup, and
b) Archive command is active, and
c) WAL contains full page writes, and
d) full_page_writes=on.
No logical log will be written to WAL in this case, and
2. During recovery, xl_tot_len check will be skipped for compressed WAL
records.

Please note that new GUC is not needed in this patch.

With this patch, compress/decompress can be developped outside the core.

I'd be very grateful if this patch can be considered again.

Best Regards;

--
-------------
Koichi Suzuki
diff -cr pgsql_org/src/backend/access/transam/xlog.c pgsql/src/backend/access/transam/xlog.c
*** pgsql_org/src/backend/access/transam/xlog.c    2007-05-02 15:56:38.000000000 +0900
--- pgsql/src/backend/access/transam/xlog.c    2007-05-07 16:30:38.000000000 +0900
***************
*** 837,842 ****
--- 837,854 ----
          return RecPtr;
      }

+     /*
+      * If online backup is not in progress and WAL archiving is active, mark
+      * backup blocks removable if any.
+      * This mark will be referenced during archiving to remove needless backup
+      * blocks in the record and compress WAL segment files.
+      */
+     if (XLogArchivingActive() && fullPageWrites &&
+             (info & XLR_BKP_BLOCK_MASK) && !Insert->forcePageWrites)
+     {
+         info |= XLR_BKP_REMOVABLE;
+     }
+
      /* Insert record header */

      record = (XLogRecord *) Insert->currpos;
***************
*** 2738,2750 ****
          blk += blen;
      }

!     /* Check that xl_tot_len agrees with our calculation */
!     if (blk != (char *) record + record->xl_tot_len)
      {
!         ereport(emode,
!                 (errmsg("incorrect total length in record at %X/%X",
!                         recptr.xlogid, recptr.xrecoff)));
!         return false;
      }

      /* Finally include the record header */
--- 2750,2778 ----
          blk += blen;
      }

!     /*
!      * If physical log has not been removed, check the length to see
!      * the following.
!      *   - No physical log existed originally,
!      *   - WAL record was not removable because it is generated during
!      *     the online backup,
!      *   - Cannot be removed because the physical log spanned in
!      *     two segments.
!      * The reason why we skip the length check on the physical log removal is
!      * that the flag XLR_SET_BKB_BLOCK(0..2) is reset to zero and it prevents
!      * the above loop to proceed blk to the end of the record.
!      */
!     if (!(record->xl_info & XLR_BKP_REMOVABLE) ||
!         record->xl_info & XLR_BKP_BLOCK_MASK)
      {
!         /* Check that xl_tot_len agrees with our calculation */
!         if (blk != (char *) record + record->xl_tot_len)
!         {
!             ereport(emode,
!                     (errmsg("incorrect total length in record at %X/%X",
!                             recptr.xlogid, recptr.xrecoff)));
!             return false;
!         }
      }

      /* Finally include the record header */
pgsql/src/backend/access/transamだけに発見: xlog.c.orig
diff -cr pgsql_org/src/include/access/xlog.h pgsql/src/include/access/xlog.h
*** pgsql_org/src/include/access/xlog.h    2007-01-06 07:19:51.000000000 +0900
--- pgsql/src/include/access/xlog.h    2007-05-07 16:30:38.000000000 +0900
***************
*** 66,73 ****
  /*
   * If we backed up any disk blocks with the XLOG record, we use flag bits in
   * xl_info to signal it.  We support backup of up to 3 disk blocks per XLOG
!  * record.    (Could support 4 if we cared to dedicate all the xl_info bits for
!  * this purpose; currently bit 0 of xl_info is unused and available.)
   */
  #define XLR_BKP_BLOCK_MASK        0x0E    /* all info bits used for bkp blocks */
  #define XLR_MAX_BKP_BLOCKS        3
--- 66,74 ----
  /*
   * If we backed up any disk blocks with the XLOG record, we use flag bits in
   * xl_info to signal it.  We support backup of up to 3 disk blocks per XLOG
!  * record.
!  * Bit 0 of xl_info is used to represent that backup blocks are not necessary
!  * in archive-log.
   */
  #define XLR_BKP_BLOCK_MASK        0x0E    /* all info bits used for bkp blocks */
  #define XLR_MAX_BKP_BLOCKS        3
***************
*** 75,80 ****
--- 76,82 ----
  #define XLR_BKP_BLOCK_1            XLR_SET_BKP_BLOCK(0)    /* 0x08 */
  #define XLR_BKP_BLOCK_2            XLR_SET_BKP_BLOCK(1)    /* 0x04 */
  #define XLR_BKP_BLOCK_3            XLR_SET_BKP_BLOCK(2)    /* 0x02 */
+ #define XLR_BKP_REMOVABLE        XLR_SET_BKP_BLOCK(3)    /* 0x01 */

  /*
   * Sometimes we log records which are out of transaction control.

Re: [PATCHES] Full page writes improvement, code update

From
Bruce Momjian
Date:
Your patch has been added to the PostgreSQL unapplied patches list at:

    http://momjian.postgresql.org/cgi-bin/pgpatches

It will be applied as soon as one of the PostgreSQL committers reviews
and approves it.

---------------------------------------------------------------------------


Koichi Suzuki wrote:
> Hi,
>
> As replied to "Patch queue triage" by Tom, here's simplified patch to
> mark WAL record as "compressable", with no increase in WAL itself.
> Compression/decompression commands will be posted separately to PG
> Foundary for further review.
>
> -------------------------------
> As suggested by Tom, I agree that WAL should not include "both" full
> page write and incremental (logical) log.   I began to examine WAL
> record format to see if incremental log can be made from full page
> writes.   It will be okay even before 8.4, if simplified patch to the
> core is accepted.   I will post simplified patch to the core as follows:
>
> 1. Mark the flag to indicate that the WAL record is compressable from
> full page writes to incremental log.  This flag will be set if
> a) It is not written during the hot backup, and
> b) Archive command is active, and
> c) WAL contains full page writes, and
> d) full_page_writes=on.
> No logical log will be written to WAL in this case, and
> 2. During recovery, xl_tot_len check will be skipped for compressed WAL
> records.
>
> Please note that new GUC is not needed in this patch.
>
> With this patch, compress/decompress can be developped outside the core.
>
> I'd be very grateful if this patch can be considered again.
>
> Best Regards;
>
> --
> -------------
> Koichi Suzuki

> diff -cr pgsql_org/src/backend/access/transam/xlog.c pgsql/src/backend/access/transam/xlog.c
> *** pgsql_org/src/backend/access/transam/xlog.c    2007-05-02 15:56:38.000000000 +0900
> --- pgsql/src/backend/access/transam/xlog.c    2007-05-07 16:30:38.000000000 +0900
> ***************
> *** 837,842 ****
> --- 837,854 ----
>           return RecPtr;
>       }
>
> +     /*
> +      * If online backup is not in progress and WAL archiving is active, mark
> +      * backup blocks removable if any.
> +      * This mark will be referenced during archiving to remove needless backup
> +      * blocks in the record and compress WAL segment files.
> +      */
> +     if (XLogArchivingActive() && fullPageWrites &&
> +             (info & XLR_BKP_BLOCK_MASK) && !Insert->forcePageWrites)
> +     {
> +         info |= XLR_BKP_REMOVABLE;
> +     }
> +
>       /* Insert record header */
>
>       record = (XLogRecord *) Insert->currpos;
> ***************
> *** 2738,2750 ****
>           blk += blen;
>       }
>
> !     /* Check that xl_tot_len agrees with our calculation */
> !     if (blk != (char *) record + record->xl_tot_len)
>       {
> !         ereport(emode,
> !                 (errmsg("incorrect total length in record at %X/%X",
> !                         recptr.xlogid, recptr.xrecoff)));
> !         return false;
>       }
>
>       /* Finally include the record header */
> --- 2750,2778 ----
>           blk += blen;
>       }
>
> !     /*
> !      * If physical log has not been removed, check the length to see
> !      * the following.
> !      *   - No physical log existed originally,
> !      *   - WAL record was not removable because it is generated during
> !      *     the online backup,
> !      *   - Cannot be removed because the physical log spanned in
> !      *     two segments.
> !      * The reason why we skip the length check on the physical log removal is
> !      * that the flag XLR_SET_BKB_BLOCK(0..2) is reset to zero and it prevents
> !      * the above loop to proceed blk to the end of the record.
> !      */
> !     if (!(record->xl_info & XLR_BKP_REMOVABLE) ||
> !         record->xl_info & XLR_BKP_BLOCK_MASK)
>       {
> !         /* Check that xl_tot_len agrees with our calculation */
> !         if (blk != (char *) record + record->xl_tot_len)
> !         {
> !             ereport(emode,
> !                     (errmsg("incorrect total length in record at %X/%X",
> !                             recptr.xlogid, recptr.xrecoff)));
> !             return false;
> !         }
>       }
>
>       /* Finally include the record header */
> pgsql/src/backend/access/transamだけに発見: xlog.c.orig
> diff -cr pgsql_org/src/include/access/xlog.h pgsql/src/include/access/xlog.h
> *** pgsql_org/src/include/access/xlog.h    2007-01-06 07:19:51.000000000 +0900
> --- pgsql/src/include/access/xlog.h    2007-05-07 16:30:38.000000000 +0900
> ***************
> *** 66,73 ****
>   /*
>    * If we backed up any disk blocks with the XLOG record, we use flag bits in
>    * xl_info to signal it.  We support backup of up to 3 disk blocks per XLOG
> !  * record.    (Could support 4 if we cared to dedicate all the xl_info bits for
> !  * this purpose; currently bit 0 of xl_info is unused and available.)
>    */
>   #define XLR_BKP_BLOCK_MASK        0x0E    /* all info bits used for bkp blocks */
>   #define XLR_MAX_BKP_BLOCKS        3
> --- 66,74 ----
>   /*
>    * If we backed up any disk blocks with the XLOG record, we use flag bits in
>    * xl_info to signal it.  We support backup of up to 3 disk blocks per XLOG
> !  * record.
> !  * Bit 0 of xl_info is used to represent that backup blocks are not necessary
> !  * in archive-log.
>    */
>   #define XLR_BKP_BLOCK_MASK        0x0E    /* all info bits used for bkp blocks */
>   #define XLR_MAX_BKP_BLOCKS        3
> ***************
> *** 75,80 ****
> --- 76,82 ----
>   #define XLR_BKP_BLOCK_1            XLR_SET_BKP_BLOCK(0)    /* 0x08 */
>   #define XLR_BKP_BLOCK_2            XLR_SET_BKP_BLOCK(1)    /* 0x04 */
>   #define XLR_BKP_BLOCK_3            XLR_SET_BKP_BLOCK(2)    /* 0x02 */
> + #define XLR_BKP_REMOVABLE        XLR_SET_BKP_BLOCK(3)    /* 0x01 */
>
>   /*
>    * Sometimes we log records which are out of transaction control.
>
>
> ---------------------------(end of broadcast)---------------------------
> TIP 9: In versions below 8.0, the planner will ignore your desire to
>        choose an index scan if your joining column's datatypes do not
>        match

--
  Bruce Momjian  <bruce@momjian.us>          http://momjian.us
  EnterpriseDB                               http://www.enterprisedb.com

  + If your life is a hard drive, Christ can be your backup. +

Re: [PATCHES] Full page writes improvement, code update

From
Tom Lane
Date:
Koichi Suzuki <suzuki.koichi@oss.ntt.co.jp> writes:
> As replied to "Patch queue triage" by Tom, here's simplified patch to
> mark WAL record as "compressable", with no increase in WAL itself.
> Compression/decompression commands will be posted separately to PG
> Foundary for further review.

Applied with some minor modifications.  I didn't like the idea of
suppressing the sanity-check on WAL record length; I think that's
fairly important.  Instead, I added a provision for an XLOG_NOOP
WAL record type that can be used to fill in the extra space.
The way I envision that working is that the compressor removes
backup blocks and converts each compressible WAL record to have the
same contents and length it would've had if written without backup
blocks.  Then, it inserts an XLOG_NOOP record with length set to
indicate the amount of extra space that needs to be chewed up --
but in the compressed version of the WAL file, XLOG_NOOP's "data
area" is not actually stored.  The decompressor need only scan
the file looking for XLOG_NOOP and insert the requisite number of
zero bytes (and maybe recompute the XLOG_NOOP's CRC, depending on
whether you want it to be valid for the short-format record in the
compressed file).  There will also be some games to be played for
WAL page boundaries, but you had to do that anyway.

            regards, tom lane

Re: [PATCHES] Full page writes improvement, code update

From
"Koichi Suzuki"
Date:
I really appreciate for the modification.

I also believe XLOG_NOOP is cool to maintains XLOG format consistent.
 I'll continue to write a code to produce incremental log record from
the full page writes as well as  too maintain CRC, XLOOG_NOOP and
other XLOG locations,    I also found that you've added information on
btree strip log records, which anables to produce corresponding
incremental logs from the full page writes.


2007/5/21, Tom Lane <tgl@sss.pgh.pa.us>:
> Koichi Suzuki <suzuki.koichi@oss.ntt.co.jp> writes:
> > As replied to "Patch queue triage" by Tom, here's simplified patch to
> > mark WAL record as "compressable", with no increase in WAL itself.
> > Compression/decompression commands will be posted separately to PG
> > Foundary for further review.
>
> Applied with some minor modifications.  I didn't like the idea of
> suppressing the sanity-check on WAL record length; I think that's
> fairly important.  Instead, I added a provision for an XLOG_NOOP
> WAL record type that can be used to fill in the extra space.
> The way I envision that working is that the compressor removes
> backup blocks and converts each compressible WAL record to have the
> same contents and length it would've had if written without backup
> blocks.  Then, it inserts an XLOG_NOOP record with length set to
> indicate the amount of extra space that needs to be chewed up --
> but in the compressed version of the WAL file, XLOG_NOOP's "data
> area" is not actually stored.  The decompressor need only scan
> the file looking for XLOG_NOOP and insert the requisite number of
> zero bytes (and maybe recompute the XLOG_NOOP's CRC, depending on
> whether you want it to be valid for the short-format record in the
> compressed file).  There will also be some games to be played for
> WAL page boundaries, but you had to do that anyway.
>
>                         regards, tom lane
>
> ---------------------------(end of broadcast)---------------------------
> TIP 3: Have you checked our extensive FAQ?
>
>                http://www.postgresql.org/docs/faq
>


--
------
Koichi Suzuki

Re: [PATCHES] Full page writes improvement, code update

From
Simon Riggs
Date:
On Tue, 2007-04-10 at 16:23 +0900, Koichi Suzuki wrote:

> Here're two patches for
>
> 1) lesslog_core.patch, patch for core, to set a mark to the log record
> to be removed in archiving,
>
> 2) lesslog_contrib.patch, patch for contrib/lesslog, pg_compresslog and
> pg_decompresslog,
>
> respectively, as asked.  I hope they work.

Koichi-san,

Earlier, I offered to document the use of pg_compresslog and
pg_decompresslog and would like to do that now.

My understanding was that we would make these programs available on
pgfoundry.org. Unfortunately, I can't find these files there, so perhaps
I misunderstood.

Do we have later versions of these programs that work with the changes
Tom committed on 20 May? Or is the code posted here the latest version?

Many thanks,

--
  Simon Riggs
  2ndQuadrant  http://www.2ndQuadrant.com