Re: vacuumdb --jobs deadlock: how to avoid pg_catalog ?

From: Adrian Klaver <adrian(dot)klaver(at)aklaver(dot)com>
To: Eduard Rozenberg <edrozenberg(at)gmail(dot)com>, pgsql-general(at)postgresql(dot)org
Subject: Re: vacuumdb --jobs deadlock: how to avoid pg_catalog ?
Date: 2020-05-14 17:35:15
Message-ID: b4d783b6-845f-f728-6124-943e6007ce01@aklaver.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-general

On 5/14/20 8:35 AM, Eduard Rozenberg wrote:
> Hello there,
>
> I'm a long-time postgres user but vacuumdb newbie :).
>
> Using vacuumdb (v12.2) with '---jobs' to vacuum a v9.6 database on
> localhost with parallel processes:
>
> $ time vacuumdb --full --echo -U postgres --jobs=8 -d mydatabase_test -p
> 5434 --password
>
> As shown below I ran into the (100%?) predictable/expected deadlock with
> sys catalog tables documented in the vacuumdb doc page
> (https://www.postgresql.org/docs/12/app-vacuumdb.html).
>
> Question: I specified "-d mydatabase_test" - why is it also doing a full
> vac on the system catalog (postgres db) which is practically guaranteed

The postgres database and the system catalogs are two separate things.

> to cause a deadlock? I don't need or want it to do a full vac on the
> postgres db as I can do that myself later if I need to, and it takes no
> time compared to full vac of my own 7 TB database that I am "extremely
> eager" to parallelize.

See:

https://www.postgresql.org/docs/12/sql-vacuum.html

"FULL

Selects “full” vacuum, which can reclaim more space, but takes much
longer and exclusively locks the table. This method also requires extra
disk space, since it writes a new copy of the table and doesn't release
the old copy until the operation is complete. Usually this should only
be used when a significant amount of space needs to be reclaimed from
within the table.
"

Pretty sure all that rewriting of tables bloats the system catalogs and
so they get a VACUUM FULL.

The question here is why is it necessary to do a VACUUM FULL in the
first place?

>
> Is my only option list all schemas.tables and feed to vacuumdb as "-t
> schema1.table1 -t schema1.table2 ........" ? That's manageable but
> unfortunate as it creates more work (having to list all tables) and adds
> additional work for automation scripts as well.
>
> I've tried searching forums/goog for past issues related to this. I did
> find an old discussion about possible enhancements to "vacuumdb --jobs"
> to help the user avoid these predictable deadlocks, where the conclusion
> was to let the user figure out a way to deal with it such as via "-t
> table -t table -t table .......".
>
> Maybe "--jobs" should only apply to the user's own databases, and when
> vacuumdb is working on system catalog tables it should apply a strictly
> serial/sequential approach to handle those tables?
>
> Thanks!
>
>
> ------------------------------------------
> output from vacuumdb's --echo
> ------------------------------------------
> ...
> VACUUM (FULL) mycompanytesting.sometable1;
> VACUUM (FULL) index.mydatabasestd_missing;
> VACUUM (FULL) mycompanytesting.sometable2;
> VACUUM (FULL) mycompanytesting.sometable3;
> VACUUM (FULL) pg_catalog.pg_attribute;
> VACUUM (FULL) mycompanydata.sometable4;
> VACUUM (FULL) pg_catalog.pg_statistic;
> VACUUM (FULL) mycompanytesting.sometable5;
> VACUUM (FULL) pg_catalog.pg_proc;
> VACUUM (FULL) pg_catalog.pg_depend;
> VACUUM (FULL) pg_catalog.pg_class;
> VACUUM (FULL) pg_catalog.pg_type;
> vacuumdb: error: vacuuming of database "mydatabase_test" failed: ERROR:
>  deadlock detected
> DETAIL:  Process 28183 waits for AccessShareLock on relation 1259 of
> database 35239378; blocked by process 28182.
> Process 28182 waits for AccessShareLock on relation 1247 of database
> 35239378; blocked by process 28183.
> HINT:  See server log for query details.
>
>
> ------------------------------
> /var/log/postgresql-9.6
> ------------------------------
> ...
> LOG:  received SIGHUP, reloading configuration files
> ERROR:  canceling autovacuum task
> CONTEXT:  automatic vacuum of table "mydatabase_test.someschema.sometable"
> ERROR:  canceling autovacuum task
> CONTEXT:  automatic vacuum of table "mydatabase_test.someschema.sometable"
> ERROR:  canceling autovacuum task
> CONTEXT:  automatic vacuum of table
> "mydatabase_test.pg_toast.pg_toast_35240266"
> LOG:  could not receive data from client: Connection timed out
> LOG:  received SIGHUP, reloading configuration files
> LOG:  received SIGHUP, reloading configuration files
> ERROR:  canceling autovacuum task
> CONTEXT:  automatic vacuum of table "mydatabase_test.someschema.sometable"
> LOG:  received SIGHUP, reloading configuration files
> LOG:  received SIGHUP, reloading configuration files
> ERROR:  canceling autovacuum task
> CONTEXT:  automatic vacuum of table
> "mydatabase_test.pg_toast.pg_toast_35240330"
> ERROR:  canceling autovacuum task
> CONTEXT:  automatic vacuum of table "mydatabase_test.someschema.sometable"
> ERROR:  deadlock detected
> DETAIL:  Process 28183 waits for AccessShareLock on relation 1259 of
> database 35239378; blocked by process 28182.
> Process 28182 waits for AccessShareLock on relation 1247 of database
> 35239378; blocked by process 28183.
> Process 28183: VACUUM (FULL) pg_catalog.pg_type;
> Process 28182: VACUUM (FULL) pg_catalog.pg_class;
> HINT:  See server log for query details.
> STATEMENT:  VACUUM (FULL) pg_catalog.pg_type;
> ERROR:  canceling statement due to user request
> STATEMENT:  VACUUM (FULL) pg_catalog.pg_class;
> ERROR:  canceling statement due to user request
> STATEMENT:  VACUUM (FULL) pg_catalog.pg_depend;
> ERROR:  canceling statement due to user request
> STATEMENT:  VACUUM (FULL) someschema.sometable
> LOG:  could not send data to client: Broken pipe
> FATAL:  connection to client lost
> LOG:  could not send data to client: Broken pipe
> FATAL:  connection to client lost
> ERROR:  canceling statement due to user request
> STATEMENT:  VACUUM (FULL) mycompanydata.wildcard_replacement_bkp;
> LOG:  could not send data to client: Broken pipe
> STATEMENT:  VACUUM (FULL) mycompanydata.wildcard_replacement_bkp;
> FATAL:  connection to client lost
> ERROR:  canceling statement due to user request
> STATEMENT:  VACUUM (FULL) pg_catalog.pg_proc;
> LOG:  could not send data to client: Broken pipe
> STATEMENT:  VACUUM (FULL) pg_catalog.pg_proc;
> FATAL:  connection to client lost
> ERROR:  canceling statement due to user request
> STATEMENT:  VACUUM (FULL) mycompanydata.cpc_tag_score2;
> LOG:  could not send data to client: Broken pipe
> STATEMENT:  VACUUM (FULL) mycompanydata.cpc_tag_score2;
> LOG:  could not send data to client: Broken pipe
> FATAL:  connection to client lost
> FATAL:  connection to client lost
> LOG:  received SIGHUP, reloading configuration files
> LOG:  received SIGHUP, reloading configuration files
>

--
Adrian Klaver
adrian(dot)klaver(at)aklaver(dot)com

In response to

Responses

Browse pgsql-general by date

  From Date Subject
Next Message Eduard Rozenberg 2020-05-14 18:27:58 Re: vacuumdb --jobs deadlock: how to avoid pg_catalog ?
Previous Message Eduard Rozenberg 2020-05-14 15:35:56 vacuumdb --jobs deadlock: how to avoid pg_catalog ?