On Mon, 6 Oct 2025 at 17:19, wenhui qiu <qiuwenhuifx@gmail.com> wrote: > I really can't agree more. Many default values are just too conservative, and the documentation doesn't provide best practices.,i think reduce to 1.x,Or add a tip in the document, providing a recommended value for different SSDs.
Did you read Tomas's email or just the subject line? I think if you're going to propose to move it in the opposite direction as to what Tomas found to be the more useful direction, then that at least warrants providing some evidence to the contrary of what Tomas has shown or stating that you think his methodology for his calculation is flawed because...
I suspect all you've done here is propagate the typical advice people give out around here. It appears to me that Tomas went to great lengths to not do that.
+1
The problem will be in estimation of the effect of cache. It can be pretty wide range.
I have a access to not too small eshop in Czech Republic (but it is not extra big) - It uses today classic stack - Java (ORM), Elastic, Postgres. The database size is cca 1.9T, shared buffers are 32GB (it handles about 10-20K logged users at one time).
The buffer cache hit ratio is 98.42%. The code is well optimized. This ratio is not calculated with file system cache.
I believe so for different applications (OLAP) or less well optimized the cache hit ratio can be much much worse.
Last year I had an experience with customers that had Postgres in clouds, and common (not extra expensive) discs are not great parameters today. It is a question if one ratio like random page cost / seq page cost can well describe dynamic throttling (or dynamic behavior of current clouds io) where customers frequently touch limits.