Re: Add the ability to limit the amount of memory that can be allocated to backends. - Mailing list pgsql-hackers
From | Andres Freund |
---|---|
Subject | Re: Add the ability to limit the amount of memory that can be allocated to backends. |
Date | |
Msg-id | 20231104041900.27pgqsknb3yl45zy@awork3.anarazel.de Whole thread Raw |
In response to | Re: Add the ability to limit the amount of memory that can be allocated to backends. (John Morris <john.morris@crunchydata.com>) |
Responses |
Re: Add the ability to limit the amount of memory that can be allocated to backends.
|
List | pgsql-hackers |
Hi, On 2023-10-31 17:11:26 +0000, John Morris wrote: > Postgres memory reservations come from multiple sources. > > * Malloc calls made by the Postgres memory allocators. > * Static shared memory created by the postmaster at server startup, > * Dynamic shared memory created by the backends. > * A fixed amount (1Mb) of “initial” memory reserved whenever a process starts up. > > Each process also maintains an accurate count of its actual memory > allocations. The process-private variable “my_memory” holds the total > allocations for that process. Since there can be no contention, each process > updates its own counters very efficiently. I think this will introduce measurable overhead in low concurrency cases and very substantial overhead / contention when there is a decent amount of concurrency. This makes all memory allocations > 1MB contend on a single atomic. Massive amount of energy have been spent writing multi-threaded allocators that have far less contention than this - the current state is to never contend on shared resources on any reasonably common path. This gives away one of the few major advantages our process model has away. The patch doesn't just introduce contention when limiting is enabled - it introduces it even when memory usage is just tracked. It makes absolutely no sense to have a single contended atomic in that case - just have a per-backend variable in shared memory that's updated. It's *WAY* cheaper to compute the overall memory usage during querying than to keep a running total in shared memory. > Pgstat now includes global memory counters. These shared memory counters > represent the sum of all reservations made by all Postgres processes. For > efficiency, these global counters are only updated when new reservations > exceed a threshold, currently 1 Mb for each process. Consequently, the > global reservation counters are approximate totals which may differ from the > actual allocation totals by up to 1 Mb per process. I see that you added them to the "cumulative" stats system - that doesn't immediately makes sense to me - what you're tracking here isn't an accumulating counter, it's something showing the current state, right? > The max_total_memory limit is checked whenever the global counters are > updated. There is no special error handling if a memory allocation exceeds > the global limit. That allocation returns a NULL for malloc style > allocations or an ENOMEM for shared memory allocations. Postgres has > existing mechanisms for dealing with out of memory conditions. I still think it's extremely unwise to do tracking of memory and limiting of memory in one patch. You should work towards and acceptable patch that just tracks memory usage in an as simple and low overhead way as possible. Then we later can build on that. > For sanity checking, pgstat now includes the pg_backend_memory_allocation > view showing memory allocations made by the backend process. This view > includes a scan of the top memory context, so it compares memory allocations > reported through pgstat with actual allocations. The two should match. Can't you just do this using the existing pg_backend_memory_contexts view? > Performance-wise, there was no measurable impact with either pgbench or a > simple “SELECT * from series” query. That seems unsurprising - allocations aren't a major part of the work there, you'd have to regress by a lot to see memory allocator changes to show a significant performance decrease. > diff --git a/src/test/regress/expected/opr_sanity.out b/src/test/regress/expected/opr_sanity.out > index 7a6f36a6a9..6c813ec465 100644 > --- a/src/test/regress/expected/opr_sanity.out > +++ b/src/test/regress/expected/opr_sanity.out > @@ -468,9 +468,11 @@ WHERE proallargtypes IS NOT NULL AND > ARRAY(SELECT proallargtypes[i] > FROM generate_series(1, array_length(proallargtypes, 1)) g(i) > WHERE proargmodes IS NULL OR proargmodes[i] IN ('i', 'b', 'v')); > - oid | proname | proargtypes | proallargtypes | proargmodes > ------+---------+-------------+----------------+------------- > -(0 rows) > + oid | proname | proargtypes | proallargtypes | proargmodes > +------+----------------------------------+-------------+---------------------------+------------------- > + 9890 | pg_stat_get_memory_reservation | | {23,23,20,20,20,20,20,20} | {i,o,o,o,o,o,o,o} > + 9891 | pg_get_backend_memory_allocation | | {23,23,20,20,20,20,20} | {i,o,o,o,o,o,o} > +(2 rows) This indicates that your pg_proc entries are broken, they need to fixed rather than allowed here. Greetings, Andres Freund
pgsql-hackers by date: