Re: best way to write large data-streams quickly?

From: Steve Atkins <steve(at)blighty(dot)com>
To: Postgres General <pgsql-general(at)postgresql(dot)org>
Subject: Re: best way to write large data-streams quickly?
Date: 2018-04-09 16:01:08
Message-ID: 1137E1BF-BEEE-4EB9-B284-2AC923C2A016@blighty.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-general


> On Apr 9, 2018, at 8:49 AM, Mark Moellering <markmoellering(at)psyberation(dot)com> wrote:
>
> Everyone,
>
> We are trying to architect a new system, which will have to take several large datastreams (total of ~200,000 parsed files per second) and place them in a database. I am trying to figure out the best way to import that sort of data into Postgres.
>
> I keep thinking i can't be the first to have this problem and there are common solutions but I can't find any. Does anyone know of some sort method, third party program, etc, that can accept data from a number of different sources, and push it into Postgres as fast as possible?

Take a look at http://ossc-db.github.io/pg_bulkload/index.html. Check the benchmarks for different situations compared to COPY.

Depending on what you're doing using custom code to parse your data and then do multiple binary COPYs in parallel may be better.

Cheers,
Steve

In response to

Responses

Browse pgsql-general by date

  From Date Subject
Next Message Peter Eisentraut 2018-04-09 17:45:08 Re: List all columns referencing an FK
Previous Message Mark Moellering 2018-04-09 15:49:10 best way to write large data-streams quickly?