From: | Alexander Antonakakis <alexis(at)maich(dot)gr> |
---|---|
To: | pgsql-general(at)postgresql(dot)org |
Subject: | Large database |
Date: | 2004-11-10 14:10:43 |
Message-ID: | 1100095843.670643@athnrd02 |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-general |
I would like to ask the more experienced users on Postgres database a
couple of questions I have on a db I manage with a lot of data. A lot of
data means something like 15.000.000 rows in a table. I will try to
describe the tables and what I will have to do on them :)
There is a table that has product data in the form of
Table product:
product_id varchar(8),
product_name text
and
product actions table:
product_id varchar(8),
flow char(1),
who int,
where int,
value float.
I will have to make sql queries in the form "select value from
product_actions where who='someone' and where='somewhere' and maybe make
also some calculations on these results. I allready have made some
indexes on these tables and a view that joins the two of them but I
would like to ask you people if someone is using such a big db and how
can I speed up things as much as it is possible on this ... these
product_actions tables exists for each year from 1988 till 2003 so this
means a lot of data...
Thanks in Advance
From | Date | Subject | |
---|---|---|---|
Next Message | Lamar Owen | 2004-11-10 14:11:25 | [SPOOFED] Nothing to see here (was:Re: I have had enough) |
Previous Message | Oliver Elphick | 2004-11-10 14:10:27 | Re: I have had enough |