Add time spent in posix_fadvise() in I/O read time ? - Mailing list pgsql-hackers

From Frédéric Yhuel
Subject Add time spent in posix_fadvise() in I/O read time ?
Date
Msg-id 3c0e7998-1001-43da-83a0-263bf79bc924@dalibo.com
Whole thread Raw
Responses Re: Add time spent in posix_fadvise() in I/O read time ?
List pgsql-hackers
Hi,

I'm currently looking at how we can integrate effective_io_concurrency 
into the planner cost model. This is actually a big overstatement, as I 
have only done a bunch of tests so far. One thing I've noticed is that 
posix_fadvise(,,POSIX_FADV_WILLNEED) isn't always non-blocking on Linux. 
As Ted Ts'o explains in this old thread[1], it blocks when the request 
queue fills up.

This leads to very surprising EXPLAIN output. In the following test, run 
on a machine with an ATA SSD (I can't observe the same on a machine with 
a nvme SSD), if we increase effective_io_concurrency from 256 to 512, 
the I/O shared read time decreases from 1148ms to 350ms, while the 
execution time remains almost the same:

CREATE UNLOGGED TABLE eic_uniform(a INT, filler TEXT);

INSERT INTO eic_uniform SELECT 10000 * random(), md5(i::text) FROM 
generate_series(1,20000000) AS T(i);

VACUUM ANALYZE eic_uniform;

CREATE INDEX ON eic_uniform(a);

set work_mem = '16MB';
set max_parallel_workers_per_gather = 0;
set track_io_timing = on;

set effective_io_concurrency = 256;

-- remove heap from postgres cache and page cache

EXPLAIN (ANALYZE, SETTINGS, BUFFERS, COSTS OFF) SELECT * FROM 
eic_uniform WHERE a < 140 AND filler > 'fffffff';
                                               QUERY PLAN 

------------------------------------------------------------------------------------------------------
  Bitmap Heap Scan on eic_uniform (actual time=2107.439..2107.440 rows=0 
loops=1)
    Recheck Cond: (a < 140)
    Filter: (filler > 'fffffff'::text)
    Rows Removed by Filter: 279648
    Heap Blocks: exact=135778
    Buffers: shared read=136024
    I/O Timings: shared read=1147.903
    ->  Bitmap Index Scan on eic_uniform_a_idx (actual 
time=41.827..41.828 rows=279648 loops=1)
          Index Cond: (a < 140)
          Buffers: shared read=246
          I/O Timings: shared read=0.568
  Settings: effective_io_concurrency = '256', 
max_parallel_workers_per_gather = '0', work_mem = '16MB'
  Planning Time: 0.197 ms
  Execution Time: 2108.553 ms


  set effective_io_concurrency = 512;

-- remove heap from postgres cache and page cache

EXPLAIN (ANALYZE, SETTINGS, BUFFERS, COSTS OFF) SELECT * FROM 
eic_uniform WHERE a < 140 AND filler > 'fffffff';
                                               QUERY PLAN 

------------------------------------------------------------------------------------------------------
  Bitmap Heap Scan on eic_uniform (actual time=2118.706..2118.707 rows=0 
loops=1)
    Recheck Cond: (a < 140)
    Filter: (filler > 'fffffff'::text)
    Rows Removed by Filter: 279648
    Heap Blocks: exact=135778
    Buffers: shared read=136024
    I/O Timings: shared read=350.014
    ->  Bitmap Index Scan on eic_uniform_a_idx (actual 
time=44.619..44.620 rows=279648 loops=1)
          Index Cond: (a < 140)
          Buffers: shared read=246
          I/O Timings: shared read=0.581
  Settings: effective_io_concurrency = '512', 
max_parallel_workers_per_gather = '0', work_mem = '16MB'
  Planning Time: 0.201 ms
  Execution Time: 2119.885 ms

Now, if we apply the attached patch (it applies to REL_17_STABLE and is 
not meant to be reviewed), it becomes:

    I/O Timings: shared wait=322.476 read=1147.903

  and:

    I/O Timings: shared wait=1030.122 read=350.014

where the "wait time" measures the time spent in posix_fadvise().

I don't know if we should do something like this... this breakdown is 
useful for my study, but probably not for most users... although it 
might be interesting to have statistics in pg_stat_io that would allow 
us to detect a bad setting of effective_io_concurrency?

But if we don't, then I think that we should at least include the time 
spent in posix_fadvise() in the I/O read time, don't you?

BTW, with the same dataset, e_i_c = 0 performs twice better than 
e_i_c = 1, which means that kernel readahead does a much better job than 
postgresql prefetching with default settings in this very specific case, 
but I think that this has already been discussed somewhere.


[1] 
https://linux.kernel.narkive.com/WcGajTwh/posix-fadvise-posix-fadv-willneed-waits-before-returning#post2


Attachment

pgsql-hackers by date:

Previous
From: Peter Eisentraut
Date:
Subject: Re: Use Python "Limited API" in PL/Python
Next
From: vignesh C
Date:
Subject: Re: Separate GUC for replication origins