Re: Avg/max size of these JSON docs in Postgres - Mailing list pgsql-general

From Tomas Vondra
Subject Re: Avg/max size of these JSON docs in Postgres
Date
Msg-id 0f5ff36a-7fdb-41d2-bf0f-0219eb0a56d9@enterprisedb.com
Whole thread Raw
In response to Re: Avg/max size of these JSON docs in Postgres  (Michael Lewis <mlewis@entrata.com>)
Responses Re: Avg/max size of these JSON docs in Postgres
List pgsql-general
On 10/12/21 21:21, Michael Lewis wrote:
> On Tue, Oct 12, 2021 at 12:51 PM Simon Riggs 
> <simon.riggs@enterprisedb.com <mailto:simon.riggs@enterprisedb.com>> wrote:
> 
>     On Tue, 12 Oct 2021 at 18:53, Michael Lewis <mlewis@entrata.com
>     <mailto:mlewis@entrata.com>> wrote:
>      >
>      > On Tue, Oct 12, 2021 at 1:43 AM Simon Riggs
>     <simon.riggs@enterprisedb.com <mailto:simon.riggs@enterprisedb.com>>
>     wrote:
>      >>
>      >> On Tue, 12 Oct 2021 at 08:14, Ram Pratap Maurya
>      >> <ram.maurya@lavainternational.in
>     <mailto:ram.maurya@lavainternational.in>> wrote:
>      >>
>      >> > Confirm what is Avg/max size of these JSON docs in Postgres.
>      >>
>      >> JSON and JSONB datatypes can both be max 1GB in size.
>      >
>      > That is per row.
> 
>     No, that is per column.
> 
> 
> Yes, sorry. My attempt at clarification only muddled things. Each column 
> within each row can be up to 1GB in size is how I understand the limit.

But you're kinda right, actually. It very much is per-tuple, because in 
various places we form tuples with all the data inline. Consider for 
example this:

   create table t (a text, b text);
   alter table t alter column a set storage extended;
   alter table t alter column b set storage extended;

   insert into t select repeat(md5('a'), 512*1024*1024/32),
                        repeat(md5('b'), 512*1024*1024/32);
   ERROR:  invalid memory alloc request size 1073741880

Clearly, both values are only 512MB (no compression). Yet it fails, 
simply because tts_virtual_copy_heap_tuple calls heap_form_tuple to form 
a tuple with all values and also the tuple header.

But this succeeds, because the values are 64B shorter, leaving enough 
space for the tuple header etc.

   insert into t select repeat(md5('a'), 512*1024*1024/32 - 2),
                        repeat(md5('b'), 512*1024*1024/32);

And you can even select the data:

   select * from t;

You can even do this:

   update t set a = repeat(md5('a'), 512*1024*1024/32);

which works, so now you have a row with two 512MB values. But then 
you'll face this:

   select * from t;
   ERROR:  out of memory
   DETAIL:  Cannot enlarge string buffer containing 536870922 bytes by
            536870912 more bytes.

because printtup() builds a huge string with all the data (and if the 
columns had compression, this would be decompressed, because it goes to 
the client).

So yeah, there's an explicit 1GB limit per value, but having rows close 
to the 1GB limit is going to cause all sorts of unpredictable and rather 
painful issues :-(


regards

-- 
Tomas Vondra
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company



pgsql-general by date:

Previous
From: Adrian Klaver
Date:
Subject: Re: Fault with initcap
Next
From: Shaozhong SHI
Date:
Subject: Re: Fault with initcap