| From: | Tom Allison <tom(at)tacocat(dot)net> |
|---|---|
| To: | pgsql-general(at)postgresql(dot)org |
| Subject: | When should I worry? |
| Date: | 2007-06-10 14:43:10 |
| Message-ID: | 466C0DFE.7040407@tacocat.net |
| Views: | Whole Thread | Raw Message | Download mbox | Resend email |
| Thread: | |
| Lists: | pgsql-general |
I've started a database that's doing wonderfully and I'm watching the tables
grow and a steady clip.
Performance is great, indexes are nice, sql costs are low. As far as I can
tell, I've done a respectable job of setting up the database, tables, sequence,
indexes...
But a little math tells me that I have one table that's particularly ugly.
This is for a total of 6 users.
If the user base gets to 100 or more, I'll be hitting a billion rows before too
long. I add about 70,000 rows per user per day. At 100 users this is 7 million
rows per day. I'll hit a billion in 142 days, call it six months for simplicity.
The table itself is small (two columns: bigint, int) but I'm wondering when I'll
start to hit a knee in performance and how I can monitor that. I know where I
work (day job) they have Oracle tables with a billion rows that just plain suck.
I don't know if a billion is bad or if the DBA's were not given the right
opportunity to make their tables work.
But if they are any indication, I'll feeling some hurt when I exceed a billion
rows. Am I going to just fold up and die in six months?
I can't really expect anyone to have an answer regarding hardware, table size,
performance speeds ... but is there some way I can either monitor for this or
estimate it before it happens?
| From | Date | Subject | |
|---|---|---|---|
| Next Message | Alexander Staubo | 2007-06-10 16:36:41 | Re: When should I worry? |
| Previous Message | Martijn van Oosterhout | 2007-06-10 13:01:49 | Re: gist index on cube column |