Re: Fatal Error : Invalid Memory alloc request size 1236252631 - Mailing list pgsql-general

From Rob Sargent
Subject Re: Fatal Error : Invalid Memory alloc request size 1236252631
Date
Msg-id 7adb97ac-5433-054d-32a0-a1046f7212da@gmail.com
Whole thread Raw
In response to Re: Fatal Error : Invalid Memory alloc request size 1236252631  (Sai Teja <saitejasaichintalapudi@gmail.com>)
List pgsql-general
On 8/17/23 07:35, Sai Teja wrote:
> Hi Team,
>
> Even I used postgreSQL Large Objects by referring this link to store 
> and retrieve large files (As bytea not working)
> https://www.postgresql.org/docs/current/largeobjects.html
>
> But even now I am unable to fetch the data at once from large objects
>
> select lo_get(oid);
>
> Here I'm getting the same error message.
>
> But if I use select data from pg_large_object where loid = 49374
> Then I can fetch the data but in page wise (data splitting into rows 
> of each size 2KB)
>
> So, here how can I fetch the data at single step rather than page by 
> page without any error.
>
> And I'm just wondering how do many applications storing huge amount of 
> data in GBs? I know that there is 1GB limit for each field set by 
> postgreSQL. If so, how to deal with these kind of situations? Would 
> like to know about this to deal with real time scenarios.
>
> We need to store large content (huge volume of data) and retrieve it. 
> Currently It is not happening due to limit of field size set by 
> postgreSQL.
>
> Would request to share your insights and suggestions on this to help 
> me for resolving this issue.
>
>
>
My first attempt at handling large payload was to use Java Selector 
directly in my app.  This worked but manually chunking the data was 
tricky.  I switched to using Tomcat and it handles large http(s) 
payloads seamlessly.





pgsql-general by date:

Previous
From: Alvaro Herrera
Date:
Subject: Re: Schema renaming cascade
Next
From: Rumpi Gravenstein
Date:
Subject: Looking for PostgreSQL Tuning Workshop Presenter