| From: | Andrew Dunstan <andrew(at)dunslane(dot)net> |
|---|---|
| To: | "Florian G(dot) Pflug" <fgp(at)phlo(dot)org> |
| Cc: | Brian Hurt <bhurt(at)janestcapital(dot)com>, Tom Lane <tgl(at)sss(dot)pgh(dot)pa(dot)us>, Postgresql-Hackers <pgsql-hackers(at)postgresql(dot)org> |
| Subject: | Re: An idea for parallelizing COPY within one backend |
| Date: | 2008-02-27 16:46:08 |
| Message-ID: | 47C593D0.6080600@dunslane.net |
| Views: | Whole Thread | Raw Message | Download mbox | Resend email |
| Thread: | |
| Lists: | pgsql-hackers |
Florian G. Pflug wrote:
>
>> Would it be possible to determine when the copy is starting that this
>> case holds, and not use the parallel parsing idea in those cases?
>
> In theory, yes. In pratice, I don't want to be the one who has to
> answer to an angry user who just suffered a major drop in COPY
> performance after adding an ENUM column to his table.
>
>
I am yet to be convinced that this is even theoretically a good path to
follow. Any sufficiently large table could probably be partitioned and
then we could use the parallelism that is being discussed for pg_restore
without any modification to the backend at all. Similar tricks could be
played by an external bulk loader for third party data sources.
cheers
andrew
| From | Date | Subject | |
|---|---|---|---|
| Next Message | Brian Hurt | 2008-02-27 16:53:04 | Re: An idea for parallelizing COPY within one backend |
| Previous Message | Tom Lane | 2008-02-27 16:41:02 | Re: An idea for parallelizing COPY within one backend |