From: | Peter Geoghegan <pg(at)heroku(dot)com> |
---|---|
To: | Simon Riggs <simon(at)2ndquadrant(dot)com> |
Cc: | Andres Freund <andres(at)2ndquadrant(dot)com>, Josh Berkus <josh(at)agliodbs(dot)com>, Greg Stark <stark(at)mit(dot)edu>, PostgreSQL-development <pgsql-hackers(at)postgresql(dot)org> |
Subject: | Re: ANALYZE sampling is too good |
Date: | 2013-12-10 19:49:12 |
Message-ID: | CAM3SWZQDDA2MrFpfCpxM05Ypr0WueejQbbvVNPcXsMKr9KHWWg@mail.gmail.com |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-hackers |
On Tue, Dec 10, 2013 at 11:23 AM, Simon Riggs <simon(at)2ndquadrant(dot)com> wrote:
> However, these things presume that we need to continue scanning most
> of the blocks of the table, which I don't think needs to be the case.
> There is a better way.
Do they? I think it's one opportunistic way of ameliorating the cost.
> Back in 2005/6, I advocated a block sampling method, as described by
> Chaudri et al (ref?)
I don't think that anyone believes that not doing block sampling is
tenable, fwiw. Clearly some type of block sampling would be preferable
for most or all purposes.
--
Peter Geoghegan
From | Date | Subject | |
---|---|---|---|
Next Message | Simon Riggs | 2013-12-10 19:54:43 | Re: ANALYZE sampling is too good |
Previous Message | Alexander Korotkov | 2013-12-10 19:44:01 | Re: GIN improvements part 1: additional information |