Re: Queue table that quickly grows causes query planner to choose poor plan

From: Justin Pryzby <pryzby(at)telsasoft(dot)com>
To: David Wheeler <dwheeler(at)dgitsystems(dot)com>
Cc: "pgsql-performance(at)postgresql(dot)org" <pgsql-performance(at)postgresql(dot)org>
Subject: Re: Queue table that quickly grows causes query planner to choose poor plan
Date: 2018-06-27 18:05:18
Message-ID: 20180627180518.GF3554@telsasoft.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-performance

On Wed, Jun 27, 2018 at 03:45:26AM +0000, David Wheeler wrote:
> Hi All,
>
> I’m having performance trouble with a particular set of queries. It goes a bit like this
>
> 1) queue table is initially empty, and very narrow (1 bigint column)
> 2) we insert ~30 million rows into queue table
> 3) we do a join with queue table to delete from another table (delete from a using queue where a.id<http://a.id> = queue.id<http://queue.id>), but postgres stats say that queue table is empty, so it uses a nested loop over all 30 million rows, taking forever

If it's within a transaction, then autovacuum couldn't begin to help until it
commits. (And if it's not, then it'll be slow on its own).

It seems to me that you can't rely on autoanalyze to finish between commiting
step 2 and beginning step 3. So you're left with options like: SET
enable_nestloop=off; or manual ANALZYE (or I guess VACUUM would be adequate to
set reltuples). Maybe you can conditionalize that: if inserted>9: ANALYZE queue.

Justin

In response to

Browse pgsql-performance by date

  From Date Subject
Next Message Tom Lane 2018-06-27 18:27:33 Re: Queue table that quickly grows causes query planner to choose poor plan
Previous Message Joseph Curtin 2018-06-27 15:53:52 Re: Bug in PostgreSQL