Re: Query on partitioned table needs memory n_partitions * work_mem - Mailing list pgsql-general

From Dimitrios Apostolou
Subject Re: Query on partitioned table needs memory n_partitions * work_mem
Date
Msg-id d26e67d3-74bc-60aa-bf24-2a8fb83efe9c@gmx.net
Whole thread Raw
In response to Re: Query on partitioned table needs memory n_partitions * work_mem  (David Rowley <dgrowleyml@gmail.com>)
Responses Re: Query on partitioned table needs memory n_partitions * work_mem
Re: Query on partitioned table needs memory n_partitions * work_mem
List pgsql-general
Thank you for the feedback.

So I've managed to reduce the query to a rather simple one:

SELECT
     workitem_n, test_executable_n,
     bool_or(test_resulttype_n IN (2,3))
FROM
     test_runs_raw
GROUP BY
     workitem_n, test_executable_n
LIMIT 10;


The TABLE test_runs_raw has 1000 partitions on RANGE(workitem_n). All the
columns are various integer types. There is an index on workitem_n.



On Thu, 11 Jul 2024, David Rowley wrote:

> On Thu, 11 Jul 2024 at 13:19, Dimitrios Apostolou <jimis@gmx.net> wrote:
>> I have a table with 1000 partitions on PostgreSQL 16.
>> I notice that a fairly complicated query of the form:
>>
>> SELECT ... GROUP BY ... LIMIT ...
>>
>> causes the postgres backend process to grow insanely very fast, and the
>> kernel OOM killer to kill it rather soon.
>> It seems it tries to allocate at least 1000 * work_mem.
>
>> ->  Append
>>    ->  HashAggregate
>>      ->  Seq Scan
>>    -> ... 1000 more hashagg+seqscans
>>
>>
>> Is this allocation pattern (workmem * n_partitions) expected under any
>> scenario? I can't find it documented.  AFAIU the backend should allocate
>> up to (depth_of_execution_plan * work_mem) (putting aside the
>> hash_mem_multiplier and the parallel workers).
>
> Not depth of execution plan. It relates to the number of nodes in the
> plan which allocate work_mem or work_mem * hash_mem_multiplier.
>
> There is some documentation in [1]:
>
> "Note that a complex query might perform several sort and hash
> operations at the same time"

The latest query is not complex at all and I don't see it doing 1000s of
operations at the same time. By "number of nodes" would you add up all
HashAggregate nodes under an Append node? Here is part of the EXPLAIN
ANALYZE output:

  Limit  (cost=0.01..28.00 rows=10 width=7) (actual time=43120.466..43292.246 rows=10 loops=1)
    Output: test_runs_raw.workitem_n, test_runs_raw.test_executable_n, (bool_or((test_runs_raw.test_resulttype_n = ANY
('{2,3}'::integer[]))))
    Buffers: shared hit=96 read=883975
    I/O Timings: shared read=16284.731
    ->  Append  (cost=0.01..3416299633.71 rows=1220556171 width=7) (actual time=42968.794..43139.855 rows=10 loops=1)
          Buffers: shared hit=96 read=883975
          I/O Timings: shared read=16284.731
          ->  HashAggregate  (cost=0.01..0.02 rows=1 width=7) (actual time=10.662..10.663 rows=0 loops=1)
                Output: test_runs_raw.workitem_n, test_runs_raw.test_executable_n,
bool_or((test_runs_raw.test_resulttype_n= ANY ('{2,3}'::integer[]))) 
                Group Key: test_runs_raw.workitem_n, test_runs_raw.test_executable_n
                Batches: 1  Memory Usage: 24kB
                ->  Seq Scan on public.test_runs_raw__part_max20k test_runs_raw  (cost=0.00..0.00 rows=1 width=8)
(actualtime=9.960..9.961 rows=0 loops=1) 
                      Output: test_runs_raw.workitem_n, test_runs_raw.test_executable_n,
test_runs_raw.test_resulttype_n
          ->  HashAggregate  (cost=0.01..0.02 rows=1 width=7) (actual time=1.913..1.914 rows=0 loops=1)
                Output: test_runs_raw_1.workitem_n, test_runs_raw_1.test_executable_n,
bool_or((test_runs_raw_1.test_resulttype_n= ANY ('{2,3}'::integer[]))) 
                Group Key: test_runs_raw_1.workitem_n, test_runs_raw_1.test_executable_n
                Batches: 1  Memory Usage: 24kB
                ->  Seq Scan on public.test_runs_raw__part_max40k test_runs_raw_1  (cost=0.00..0.00 rows=1 width=8)
(actualtime=1.031..1.031 rows=0 loops=1) 
                      Output: test_runs_raw_1.workitem_n, test_runs_raw_1.test_executable_n,
test_runs_raw_1.test_resulttype_n
[ .... 1000s of similar HashAggregate nodes ... ]
Settings: temp_buffers = '32MB', work_mem = '32MB', effective_io_concurrency = '300', max_parallel_workers_per_gather =
'0',enable_hashjoin = 'off', enable_partitionwise_join = 'on', enable_partitionwise_aggregate = 'on', random_page_cost
='1.1', effective_cache_size = '6GB', from_collapse_limit = '24', join_collapse_limit = '24' 
  Planning:
    Buffers: shared hit=377
  Planning Time: 1503.800 ms
  Execution Time: 56515.185 ms
(5382 rows)

Memory usage on each HashAggregate is logged as 24KB (many HashAggregates
are missing that info though), I guess the EXPLAIN output is missing some
important part of the allocations here since I'm seeing MBs of allocations
per node.

I can't help but see this as a bug. I see many issues:

* postgres is not reading from partitions in parallel, but one after the
   other. It shouldn't need all this memory simultaneously.

* The memory is unnecessarily allocated early on, before any partitions
   are actually aggregated. I know this because I/O is slow on this device
   and the table sizes are huge, it's simply not possible that postgres
   went through all partitions and blew up the memory. That would take
   hours, but the OOM happens seconds after I start the query.

* The memory is not only allocated by the planner, but it's actually
   accessed. Libc's malloc() has no problem allocating gigabytes more than
   what I have available, growing the VSZ memory size without swapping out
   a single byte. That's because my kernel was set to overcommit.  The fact
   that postgres grows its RSS memory too and swaps out other processes and
   finally gets killed, is because it actually writes to that memory!
   I wonder what it writes, if it hasn't read any of the data on the
   partitions. :-)

For your consideration, here is also a memory dump from when the backend
consumes all the memory (work_mem increased to 256MB), and I've disabled
overcommit in the kernel:


TopMemoryContext: 573520 total in 8 blocks; 214096 free (27 chunks); 359424 used
   TopTransactionContext: 8192 total in 1 blocks; 7752 free (1 chunks); 440 used
   Record information cache: 8192 total in 1 blocks; 1616 free (0 chunks); 6576 used
   RegexpCacheMemoryContext: 1024 total in 1 blocks; 760 free (0 chunks); 264 used
     RegexpMemoryContext: 44032 total in 5 blocks; 36400 free (17 chunks); 7632 used: ^(test_runs_raw)$
   Btree proof lookup cache: 8192 total in 1 blocks; 592 free (0 chunks); 7600 used
   Operator lookup cache: 24576 total in 2 blocks; 10792 free (3 chunks); 13784 used
   TableSpace cache: 8192 total in 1 blocks; 2128 free (0 chunks); 6064 used
   Type information cache: 24368 total in 2 blocks; 2648 free (0 chunks); 21720 used
   RowDescriptionContext: 8192 total in 1 blocks; 6896 free (0 chunks); 1296 used
   MessageContext: 16777216 total in 12 blocks; 2924704 free (11 chunks); 13852512 used
     partition directory: 8192 total in 1 blocks; 592 free (0 chunks); 7600 used
   Operator class cache: 8192 total in 1 blocks; 592 free (0 chunks); 7600 used
   smgr relation table: 524288 total in 7 blocks; 231192 free (28 chunks); 293096 used
   PgStat Shared Ref Hash: 50224 total in 2 blocks; 688 free (0 chunks); 49536 used
   PgStat Shared Ref: 90112 total in 14 blocks; 3648 free (14 chunks); 86464 used
   PgStat Pending: 294912 total in 39 blocks; 193864 free (788 chunks); 101048 used
   TransactionAbortContext: 32768 total in 1 blocks; 32504 free (0 chunks); 264 used
   Portal hash: 8192 total in 1 blocks; 592 free (0 chunks); 7600 used
   TopPortalContext: 8192 total in 1 blocks; 7664 free (0 chunks); 528 used
     PortalContext: 1024 total in 1 blocks; 624 free (0 chunks); 400 used: <unnamed>
       ExecutorState: 8462432 total in 13 blocks; 3316568 free (9 chunks); 5145864 used
         HashAgg meta context: 8192 total in 1 blocks; 7720 free (0 chunks); 472 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         HashAgg meta context: 100671536 total in 2 blocks; 5528 free (0 chunks); 100666008 used
           ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         HashAgg meta context: 201334832 total in 2 blocks; 5528 free (0 chunks); 201329304 used
           ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
         ExprContext: 8192 total in 1 blocks; 7928 free (0 chunks); 264 used
[...]
         2907 more child contexts containing 13752609344 total in 3787 blocks; 25715912 free (0 chunks); 13726893432
used
-- NOTE: the total number above is the culprit for the OOM.

   Relcache by OID: 131072 total in 5 blocks; 38744 free (13 chunks); 92328 used
   CacheMemoryContext: 47147632 total in 599 blocks; 7315600 free (2 chunks); 39832032 used
     relation rules: 16384 total in 5 blocks; 2272 free (1 chunks); 14112 used: pg_settings
     index info: 1024 total in 1 blocks; 0 free (0 chunks); 1024 used: pg_trigger_oid_index
     index info: 1024 total in 1 blocks; 0 free (0 chunks); 1024 used: pg_trigger_tgconstraint_index
     index info: 1024 total in 1 blocks; 40 free (0 chunks); 984 used: pg_publication_rel_prpubid_index
     index info: 2048 total in 2 blocks; 728 free (2 chunks); 1320 used: pg_policy_polrelid_polname_index
     index info: 1024 total in 1 blocks; 72 free (0 chunks); 952 used: pg_policy_oid_index
     relation rules: 16384 total in 5 blocks; 7144 free (1 chunks); 9240 used: pg_roles
     index info: 2048 total in 2 blocks; 544 free (1 chunks); 1504 used: pg_inherits_relid_seqno_index
     index info: 1024 total in 1 blocks; 0 free (0 chunks); 1024 used: pg_tablespace_spcname_index
     index info: 1024 total in 1 blocks; 0 free (0 chunks); 1024 used: pg_constraint_contypid_index
     index info: 2048 total in 2 blocks; 688 free (1 chunks); 1360 used: pg_constraint_conname_nsp_index
     index info: 2048 total in 2 blocks; 912 free (0 chunks); 1136 used: pg_constraint_conparentid_index
     index info: 1024 total in 1 blocks; 0 free (0 chunks); 1024 used: pg_attrdef_oid_index
     index info: 2048 total in 2 blocks; 688 free (1 chunks); 1360 used: pg_attrdef_adrelid_adnum_index
     index info: 1024 total in 1 blocks; 16 free (0 chunks); 1008 used: pg_am_oid_index
     index info: 1024 total in 1 blocks; 0 free (0 chunks); 1024 used: pg_am_name_index
     index info: 1024 total in 1 blocks; 0 free (0 chunks); 1024 used: pg_rewrite_oid_index
     CachedPlanSource: 4096 total in 3 blocks; 688 free (0 chunks); 3408 used: SELECT * FROM pg_catalog.pg_rewrite
WHEREev_class = $1 AND rulename = $2 
       CachedPlanQuery: 8192 total in 4 blocks; 4016 free (1 chunks); 4176 used
     SPI Plan: 1024 total in 1 blocks; 600 free (0 chunks); 424 used
     index info: 2048 total in 2 blocks; 656 free (2 chunks); 1392 used: pg_class_tblspc_relfilenode_index
     index info: 2048 total in 2 blocks; 616 free (1 chunks); 1432 used: pg_toast_2619_index
     index info: 2048 total in 2 blocks; 728 free (2 chunks); 1320 used: test_runs_raw__part_max20000k_pkey
[...]
     1017 more child contexts containing 2142696 total in 2014 blocks; 684544 free (1928 chunks); 1458152 used
   WAL record construction: 50200 total in 2 blocks; 6376 free (0 chunks); 43824 used
   PrivateRefCount: 8192 total in 1 blocks; 2648 free (0 chunks); 5544 used
   MdSmgr: 131072 total in 5 blocks; 22552 free (1 chunks); 108520 used
   LOCALLOCK hash: 262144 total in 6 blocks; 59376 free (23 chunks); 202768 used
   GUCMemoryContext: 24576 total in 2 blocks; 9440 free (4 chunks); 15136 used
     GUC hash table: 32768 total in 3 blocks; 12704 free (5 chunks); 20064 used
   Timezones: 104112 total in 2 blocks; 2648 free (0 chunks); 101464 used
   ErrorContext: 8192 total in 1 blocks; 7928 free (2 chunks); 264 used Grand total: 16246728816 bytes in 6875 blocks;
41833616free (3045 chunks); 16204895200 used 



>
> Also, see the warning about execution time memory in [2].
>
>> NOTE: after having written the above message, it occured to me that I have
>> enable_partitionwise_aggregate=on. And Turning it off fixes the issue and
>> makes the query faster too! Expected behaviour or bug?
>
> enable_partitionwise_aggregate=on causes this query to perform an
> aggregate per partition.  If your GROUP BY clause values are
> distributed evenly throughout all partitions then you might find it's
> not much slower to execute the query with
> enable_partitionwise_aggregate=off.

The GROUP BY first-clause values (workitem_n) are not evenly distributed,
in fact the partitions are split by that. The rest indeed are.

Thank you, I'll turn off this flag globally to avoid such behaviour.

>
> It's understandable that how PostgreSQL uses work_mem isn't ideal
> here, but unfortunately, that's the state of affairs, currently. You
> might want to reconsider your enable_partitionwise_aggregate setting
> and/or how many partitions you have.

Having wasted long time in that, the minimum I can do is submit a
documentation patch. At enable_partitionwise_aggregate someting like
"WARNING it can increase the memory usage by at least
n_partitions * work_mem". How do I move on for such a patch? Pointers
would be appreciated. :-)


Thank you,
Dimitris


> [1] https://www.postgresql.org/docs/current/runtime-config-resource.html
> [2] https://www.postgresql.org/docs/current/ddl-partitioning.html#DDL-PARTITIONING-DECLARATIVE-BEST-PRACTICES
>



pgsql-general by date:

Previous
From: Alvaro Herrera
Date:
Subject: Re: Dropping column from big table
Next
From: Tom Lane
Date:
Subject: Re: Query on partitioned table needs memory n_partitions * work_mem