From: | Otandeka Simon Peter <sotandeka(at)gmail(dot)com> |
---|---|
To: | Vincenzo Romano <vincenzo(dot)romano(at)notorand(dot)it> |
Cc: | P Kishor <punk(dot)kish(at)gmail(dot)com>, jd(at)commandprompt(dot)com, Bill Thoen <bthoen(at)gisnet(dot)com>, pgsql-general General <pgsql-general(at)postgresql(dot)org> |
Subject: | Re: How Big is Too Big for Tables? |
Date: | 2010-07-28 18:30:04 |
Message-ID: | AANLkTik8P8JAXojUkEr87-93xPZ3O30ic=Qr-M64-Ro_@mail.gmail.com |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-general |
There are Postgres Enterprise solutions available although I think they are
commercial. You may want to take a look and see if they can be helpful to
you.
On Wed, Jul 28, 2010 at 8:44 PM, Vincenzo Romano <
vincenzo(dot)romano(at)notorand(dot)it> wrote:
> 2010/7/28 P Kishor <punk(dot)kish(at)gmail(dot)com>:
> ...
> > Two. Partitioning is not the perfect solution. My database will
> > ultimately have about 13 million rows per day (it is daily data) for
> > about 25 years. So, I need either --
> >
> > - One big table with 25 * 365 * 13 million rows. Completely undoable.
> > - 25 yearly tables with 365 * 13 million rows each. Still a huge
> > chore, very slow queries.
> > - 25 * 365 tables with 13 million rows each. More doable, but
> > partitioning doesn't work.
> >
> > Three. At least, in my case, the overhead is too much. My data are
> > single bytes, but the smallest data type in Pg is smallint (2 bytes).
> > That, plus the per row overhead adds to a fair amount of overhead.
> >
> > I haven't yet given up on storing this specific dataset in Pg, but am
> > reconsidering. It is all readonly data, so flat files might be better
> > for me.
> >
> > In other words, Pg is great, but do tests, benchmark, research before
> > committing to a strategy. Of course, since you are storing geometries,
> > Pg is a natural choice for you. My data are not geometries, so I can
> > explore alternatives for it, while keeping my geographic data in Pg.
>
> That recalls me an old inquiry of mine on the list about "enterprise
> grade" (or whatever you want to call it) solutions.
> That means, "really lots of rows" or, alternatively "really lots of tables
> in
> the hierarchy" or, again, "really lots of partial indexes".
>
> Partitioning is not going to work probably because coping with
> thousands of tables in a hierarchy would hit against some "linear"
> algorithm inside the query planner, even with constraint exclusion.
>
> Maybe "multilevel" hierarchy (let's say partitioning by months (12)
> on the first level *and* by day (28,29,30 or 31) on the second one)
> would do the magics, but here the DDL would be quite killing,
> even with some PL/PGSQL helper function.
>
> The "linearity" of the index selection killed the performances also in
> the "really lots of partial indexes" approach.
>
> --
> NotOrAnd Information Technologies
> Vincenzo Romano
> --
> NON QVIETIS MARIBVS NAVTA PERITVS
>
> --
> Sent via pgsql-general mailing list (pgsql-general(at)postgresql(dot)org)
> To make changes to your subscription:
> http://www.postgresql.org/mailpref/pgsql-general
>
From | Date | Subject | |
---|---|---|---|
Next Message | Stephen Frost | 2010-07-28 18:38:26 | Re: How Big is Too Big for Tables? |
Previous Message | David Fetter | 2010-07-28 18:23:36 | Re: order in which rules are executed |