From: | hubert depesz lubaczewski <depesz(at)depesz(dot)com> |
---|---|
To: | Karthik Krishnakumar <karthikk(at)zohocorp(dot)com> |
Cc: | pgsql-admin <pgsql-admin(at)lists(dot)postgresql(dot)org> |
Subject: | Re: Max tables in a cluster? |
Date: | 2023-04-05 10:21:24 |
Message-ID: | ZC1LpCz1p6hp0HXO@depesz.com |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-admin |
On Wed, Apr 05, 2023 at 12:37:41PM +0530, Karthik Krishnakumar wrote:
> Is there a limit to number of tables that a cluster can support?
Well, yes.
For starters: https://www.postgresql.org/docs/current/limits.html
But more important is: with large number of tables certain operations
become VERY slow. Specifically getting schema dump. Which is used, for
example, by pg_upgrade.
In my experience hitting 1 million rows in pg_class made it
realistically impossible to use pg_dump. But there have been
many improvements to pg_dump over the years
(https://why-upgrade.depesz.com/show?from=9.5&to=15.2&keywords=pg_dump)
to it's possible that the problem is gone now.
Try taking `pg_dump -s` of the db, and see how long it takes...
Best regards,
depesz
From | Date | Subject | |
---|---|---|---|
Next Message | Norbert Poellmann | 2023-04-05 10:37:24 | Re: Barman and Hot Standby |
Previous Message | Ron | 2023-04-05 07:20:50 | Re: Max tables in a cluster? |