From: | Andy Colson <andy(at)squeakycode(dot)net> |
---|---|
To: | pgsql-general(at)postgresql(dot)org |
Subject: | Re: Processor usage/tuning question |
Date: | 2014-10-04 02:28:43 |
Message-ID: | 542F5B5B.1010406@squeakycode.net |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-general |
On 10/03/2014 04:40 PM, Alan Hodgson wrote:
> On Friday, October 03, 2014 11:24:31 AM Israel Brewster wrote:
>> I have a Postgresql 9.3.5 server running on CentOS 6.5. In looking at some
>> stats today, I saw that it was handling about 4-5 transactions/second
>> (according to the SELECT sum(xact_commit+xact_rollback) FROM
>> pg_stat_database; query), and an instance of the postmaster process was
>> consistently showing 40%-80% utilization to handle this. I didn't think
>> anything of that (the machine has plenty of capacity) until I mentioned it
>> to a friend of mine, who said that utilization level seemed high for that
>> many transactions. So if that level of utilization IS high, what might I
>> need to tune to bring it down to a more reasonable level?
>>
>
> You probably have some read queries not properly indexed that are sequentially
> scanning that 1.2 million row table over and over again. Enable slow query
> logging and see what's going on.
>
>
>
Yep, do that... and then:
https://wiki.postgresql.org/wiki/Slow_Query_Questions
-Andy
From | Date | Subject | |
---|---|---|---|
Next Message | Tim Mickelson | 2014-10-04 05:45:43 | Re: Really strange foreign key constraint problem blocking delete |
Previous Message | Adrian Klaver | 2014-10-03 23:15:58 | Re: How to find greatest record before known values fast |