Re: best way to write large data-streams quickly? - Mailing list pgsql-general

From Steve Atkins
Subject Re: best way to write large data-streams quickly?
Date
Msg-id 1137E1BF-BEEE-4EB9-B284-2AC923C2A016@blighty.com
Whole thread Raw
In response to best way to write large data-streams quickly?  (Mark Moellering <markmoellering@psyberation.com>)
Responses Re: best way to write large data-streams quickly?
List pgsql-general
> On Apr 9, 2018, at 8:49 AM, Mark Moellering <markmoellering@psyberation.com> wrote:
>
> Everyone,
>
> We are trying to architect a new system, which will have to take several large datastreams (total of ~200,000 parsed
filesper second) and place them in a database.  I am trying to figure out the best way to import that sort of data into
Postgres.  
>
> I keep thinking i can't be the first to have this problem and there are common solutions but I can't find any.  Does
anyoneknow of some sort method, third party program, etc, that can accept data from a number of different sources, and
pushit into Postgres as fast as possible? 

Take a look at http://ossc-db.github.io/pg_bulkload/index.html. Check the benchmarks for different situations compared
toCOPY. 

Depending on what you're doing using custom code to parse your data and then do multiple binary COPYs in parallel may
bebetter. 

Cheers,
  Steve



pgsql-general by date:

Previous
From: Mark Moellering
Date:
Subject: best way to write large data-streams quickly?
Next
From: Peter Eisentraut
Date:
Subject: Re: List all columns referencing an FK