From: | Pavel Stehule <pavel(dot)stehule(at)gmail(dot)com> |
---|---|
To: | Mike Sofen <msofen(at)runbox(dot)com> |
Cc: | "pgsql-general(at)postgresql(dot)org >> PG-General Mailing List" <pgsql-general(at)postgresql(dot)org> |
Subject: | Re: how to monitor the progress of really large bulk operations? |
Date: | 2016-09-28 04:13:48 |
Message-ID: | CAFj8pRDff+Uc__n0a46kv8Nw1FYFEBpwetq4M=Tpf=s_J6anuA@mail.gmail.com |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-general |
Hi
2016-09-27 23:03 GMT+02:00 Mike Sofen <msofen(at)runbox(dot)com>:
> Hi gang,
>
>
>
> On PG 9.5.1, linux, I’m running some large ETL operations, migrate data
> from a legacy mysql system into PG, upwards of 250m rows in a transaction
> (it’s on a big box). It’s always a 2 step operation – extract raw mysql
> data and pull it to the target big box into staging tables that match the
> source, the second step being read the landed dataset and transform it into
> the final formats, linking to newly generated ids, compressing big subsets
> into jsonb documents, etc.
>
>
>
> While I could break it into smaller chunks, it hasn’t been necessary, and
> it doesn’t eliminate my need: how to view the state of a transaction in
> flight, seeing how many rows have been read or inserted (possible for a
> transaction in flight?), memory allocations across the various PG
> processes, etc.
>
>
>
> Possible or a hallucination?
>
>
>
> Mike Sofen (Synthetic Genomics)
>
some years ago I used a trick
http://okbob.blogspot.cz/2014/09/nice-unix-filter-pv.html#links
Regards
Pavel
From | Date | Subject | |
---|---|---|---|
Next Message | Pavel Stehule | 2016-09-28 04:17:39 | Re: how to monitor the progress of really large bulk operations? |
Previous Message | Tom Lane | 2016-09-27 22:48:19 | Re: Re: [GENERAL] inconsistent behaviour of set-returning functions in sub-query with random() |