| From: | bricklen <bricklen(at)gmail(dot)com> |
|---|---|
| To: | Giuseppe Broccolo <giuseppe(dot)broccolo(at)2ndquadrant(dot)it> |
| Cc: | "pgsql-admin(at)postgresql(dot)org" <pgsql-admin(at)postgresql(dot)org> |
| Subject: | Re: PostgreSQL 9.2 - pg_dump out of memory when backuping a database with 300000000 large objects |
| Date: | 2013-10-01 12:30:18 |
| Message-ID: | CAGrpgQ_safsytHcJyBwo2fT6Eu01=hJwjiZ2juac1vJQRqCjfg@mail.gmail.com |
| Views: | Whole Thread | Raw Message | Download mbox | Resend email |
| Thread: | |
| Lists: | pgsql-admin pgsql-sql |
On Tue, Oct 1, 2013 at 4:01 AM, Giuseppe Broccolo <
giuseppe(dot)broccolo(at)2ndquadrant(dot)it> wrote:
> Maybe you can performe your database changing some parameters properly:
>
> max_connections = 500 # (change requires restart)
>>
> Set it to 100, the highest value supported by PostgreSQL
>
Surely you mean that max_connections = 100 is the *default* ?
| From | Date | Subject | |
|---|---|---|---|
| Next Message | Magnus Hagander | 2013-10-01 12:49:33 | Re: PostgreSQL 9.2 - pg_dump out of memory when backuping a database with 300000000 large objects |
| Previous Message | Sergey Klochkov | 2013-10-01 11:12:18 | Re: PostgreSQL 9.2 - pg_dump out of memory when backuping a database with 300000000 large objects |
| From | Date | Subject | |
|---|---|---|---|
| Next Message | Magnus Hagander | 2013-10-01 12:49:33 | Re: PostgreSQL 9.2 - pg_dump out of memory when backuping a database with 300000000 large objects |
| Previous Message | Sergey Klochkov | 2013-10-01 11:12:18 | Re: PostgreSQL 9.2 - pg_dump out of memory when backuping a database with 300000000 large objects |