From: | Ajin Cherian <itsajin(at)gmail(dot)com> |
---|---|
To: | Amit Kapila <amit(dot)kapila16(at)gmail(dot)com> |
Cc: | Dilip Kumar <dilipbalaut(at)gmail(dot)com>, Erik Rijkers <er(at)xs4all(dot)nl>, Kuntal Ghosh <kuntalghosh(dot)2007(at)gmail(dot)com>, Tomas Vondra <tomas(dot)vondra(at)2ndquadrant(dot)com>, Michael Paquier <michael(at)paquier(dot)xyz>, Peter Eisentraut <peter(dot)eisentraut(at)2ndquadrant(dot)com>, PostgreSQL Hackers <pgsql-hackers(at)postgresql(dot)org> |
Subject: | Re: PATCH: logical_work_mem and logical streaming of large in-progress transactions |
Date: | 2020-07-10 03:51:08 |
Message-ID: | CAFPTHDZUQTa8oOthwLOpFwQP128mJOYYTbxtyV6D7Ejc60fNGw@mail.gmail.com |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-hackers |
On Thu, Jul 9, 2020 at 1:30 PM Amit Kapila <amit(dot)kapila16(at)gmail(dot)com> wrote:
>
> > I think if the GUC is set then maybe we can bypass this check so that
> > it can try to stream every single change?
> >
>
> Yeah and probably we need to do something for the check "while
> (rb->size >= logical_decoding_work_mem * 1024L)" as well.
>
>
> I have made this change, as discussed, the regression tests seem to run
fine. I have added a debug that records the streaming for each transaction
number. I also had to bypass certain asserts
in ReorderBufferLargestTopTXN() as now we are going through the entire list
of transactions and not just picking the biggest transaction .
regards,
Ajin
Fujitsu Australia
Attachment | Content-Type | Size |
---|---|---|
v31-0015-TEST-guc-always-streaming-logical.patch | application/octet-stream | 3.0 KB |
From | Date | Subject | |
---|---|---|---|
Next Message | Zhenghua Lyu | 2020-07-10 04:23:04 | distribute_restrictinfo_to_rels if restrictinfo contains volatile functions |
Previous Message | Amit Langote | 2020-07-10 03:21:12 | Re: [PATCH] Performance Improvement For Copy From Binary Files |