From: | "Brent Wood" <b(dot)wood(at)niwa(dot)co(dot)nz> |
---|---|
To: | <a_wronski(at)gazeta(dot)pl>, <pgsql-general(at)postgresql(dot)org> |
Subject: | Re: 10 TB database |
Date: | 2009-06-15 21:00:04 |
Message-ID: | 4A375F150200007B0001BA54@gwia1.ham.niwa.co.nz |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-general |
Hi Artur,
Some general comments:
I'd look at partitioning and tablespaces to better manage the files where the data is stored, but also look at some efficiently parallelised disks behind the filesystems. You might also look at optimising the filesystem &OS parameters to increase efficiency as well, so it is a mix of hardware/OS/filesystem & db setup to optimise for such a situation.
For data retrieval, clustered indexes may help, but as this requires a physical reordering of the data on disk, it may be impractical.
Cheers,
Brent Wood
Brent Wood
DBA/GIS consultant
NIWA, Wellington
New Zealand
>>> Artur <a_wronski(at)gazeta(dot)pl> 06/16/09 3:30 AM >>>
Hi!
We are thinking to create some stocks related search engine.
It is experimental project just for fun.
The problem is that we expect to have more than 250 GB of data every month.
This data would be in two tables. About 50.000.000 new rows every month.
We want to have access to all the date mostly for generating user
requesting reports (aggregating).
We would have about 10TB of data in three years.
Do you think is it possible to build this with postgresql and have any
idea how to start? :)
Thanks in advance,
Artur
--
Sent via pgsql-general mailing list (pgsql-general(at)postgresql(dot)org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general
NIWA is the trading name of the National Institute of Water & Atmospheric Research Ltd.
From | Date | Subject | |
---|---|---|---|
Next Message | David Fetter | 2009-06-15 21:02:46 | Re: Custom Fields Database Architecture |
Previous Message | Stefan Keller | 2009-06-15 20:37:04 | Re: Custom Fields Database Architecture |