Hello all.
I am managing a large database with lots of transactions in different
tables.
The largest tables have around 5-6 millions tuples and around 50000-60000
inserts and maybe 20000 updates pr day.
While the smalest tables have only a few tuples and a few updates /inserts
pr day. In addition we have small tables with many updates/inserts. So what
I am saying is that there is all kinds of tables and uses of tables in our
database.
This, I think, makes it difficult to set up pg_autovacuum. I am now running
vacuum jobs on different tables in cron.
What things should I consider when setting but base and threshold values in
pg_autovacuum? Since the running of vacuum and analyze is relative to the
table size, as it must be, I think it is difficult to cover all tables..
Are there anyone who have some thoughts around this?
Regards
Rune