Re: Out of memory error

From: Tom Lane <tgl(at)sss(dot)pgh(dot)pa(dot)us>
To: aditya desai <admad123(at)gmail(dot)com>
Cc: Pgsql Performance <pgsql-performance(at)lists(dot)postgresql(dot)org>
Subject: Re: Out of memory error
Date: 2021-11-24 05:52:38
Message-ID: 97240.1637733158@sss.pgh.pa.us
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-performance

aditya desai <admad123(at)gmail(dot)com> writes:
> In a trigger function I am creating a temp table . When an update on a
> table is executed for say 10k rows. I get the below error.

> ERROR: out of shared memory
> HINT:You might need to increase max_locks_per_transaction
> CONTEXT: SQL Statement "created temp table changedinfo(colName
> varchar(100), oldValue varchar(4000), newValue varchar(4000)

[ raised eyebrow ... ] If you are concerned about performance,
I'd start by not creating a temp table per row of the outer update.
That's costing probably 100x to 1000x as much as the row update itself.

regards, tom lane

In response to

Responses

Browse pgsql-performance by date

  From Date Subject
Next Message aditya desai 2021-11-24 06:25:50 Re: Out of memory error
Previous Message aditya desai 2021-11-24 05:27:37 Out of memory error