When should I worry?

From: Tom Allison <tom(at)tacocat(dot)net>
To: pgsql-general(at)postgresql(dot)org
Subject: When should I worry?
Date: 2007-06-10 14:43:10
Message-ID: 466C0DFE.7040407@tacocat.net
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-general

I've started a database that's doing wonderfully and I'm watching the tables
grow and a steady clip.

Performance is great, indexes are nice, sql costs are low. As far as I can
tell, I've done a respectable job of setting up the database, tables, sequence,
indexes...

But a little math tells me that I have one table that's particularly ugly.

This is for a total of 6 users.

If the user base gets to 100 or more, I'll be hitting a billion rows before too
long. I add about 70,000 rows per user per day. At 100 users this is 7 million
rows per day. I'll hit a billion in 142 days, call it six months for simplicity.

The table itself is small (two columns: bigint, int) but I'm wondering when I'll
start to hit a knee in performance and how I can monitor that. I know where I
work (day job) they have Oracle tables with a billion rows that just plain suck.
I don't know if a billion is bad or if the DBA's were not given the right
opportunity to make their tables work.

But if they are any indication, I'll feeling some hurt when I exceed a billion
rows. Am I going to just fold up and die in six months?

I can't really expect anyone to have an answer regarding hardware, table size,
performance speeds ... but is there some way I can either monitor for this or
estimate it before it happens?

Responses

Browse pgsql-general by date

  From Date Subject
Next Message Alexander Staubo 2007-06-10 16:36:41 Re: When should I worry?
Previous Message Martijn van Oosterhout 2007-06-10 13:01:49 Re: gist index on cube column