From: | "Meetesh Karia" <meetesh(dot)karia(at)gmail(dot)com> |
---|---|
To: | "Craig A(dot) James" <cjames(at)modgraph-usa(dot)com> |
Cc: | pgsql-performance(at)postgresql(dot)org, "Jean-David Dahan" <jd(dot)dahan(at)gmail(dot)com> |
Subject: | Re: Query hanging/not finishing inconsistently |
Date: | 2006-06-20 17:25:19 |
Message-ID: | fc5b04ca0606201025g7d83548fj7cf0747f0a62dffe@mail.gmail.com |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-performance |
Hi all,
I just saw another email on the mailing list to this effect as well. We
recently updated the kernel versions on our machines to the latest stable
versions (which contained both HyperThreading and IO bug fixes) and we
updated Postgres to version 8.0.8. We thought we were in the clear when we
didn't encounter a hang for 6+ days. But, once again we ran into the same
situation where a query that normally executes in ~15ms wouldn't finish. As
before, there were no ungranted locks and threads weren't waiting on a
lock. I attached gdb to one of the stuck postgres processes and got the
following stack trace:
#0 0x008967a2 in _dl_sysinfo_int80 () from /lib/ld-linux.so.2
#1 0x00977e5b in semop () from /lib/tls/libc.so.6
#2 0x08167298 in PGSemaphoreLock ()
#3 0x0818bcb5 in LWLockAcquire ()
#4 0x080a47f5 in SimpleLruWritePage ()
#5 0x080a48ad in SimpleLruReadPage ()
#6 0x080a519a in SubTransGetParent ()
#7 0x080a51f2 in SubTransGetTopmostTransaction ()
#8 0x0821371c in HeapTupleSatisfiesSnapshot ()
#9 0x080822a2 in heap_release_fetch ()
#10 0x080880fb in index_getnext ()
#11 0x08128507 in ExecReScanHashJoin ()
#12 0x08122a09 in ExecScan ()
#13 0x081287f9 in ExecIndexScan ()
#14 0x0811dfdd in ExecProcNode ()
#15 0x0812a49f in ExecNestLoop ()
#16 0x0811df9d in ExecProcNode ()
#17 0x0812b74d in ExecSort ()
#18 0x0811df5d in ExecProcNode ()
#19 0x0812b941 in ExecUnique ()
#20 0x0811df2c in ExecProcNode ()
#21 0x0811ce18 in ExecutorRun ()
#22 0x081947ec in PortalSetResultFormat ()
#23 0x08194df4 in PortalRun ()
#24 0x08192ef7 in PostgresMain ()
#25 0x08169780 in ClosePostmasterPorts ()
#26 0x0816b0ae in PostmasterMain ()
#27 0x0813a5a6 in main ()
We then upgraded glibc to 2.3.4-2.19 but we encountered the problem within a
day. Our latest attempt at isolating the problem has been to reboot the
machine with a 'noht' kernel param. The machine has been up for 1 day,
13:18 since then and we haven't seen the problem yet.
Has anyone been able to solve this problem?
Thanks,
Meetesh
On 5/22/06, Meetesh Karia <meetesh(dot)karia(at)gmail(dot)com > wrote:
>
> Hi Craig,
>
> Thanks for your response. This did start recently and it wasn't after a
> kernel update, but it was after we moved the db from Machine B to Machine A
> (which have slightly different kernel versions). However, the problem took
> about a week to show up after we moved from one machine to the other.
> Unfortunately, the problem only reappears after 15 mins once it occurs the
> first time. If it occurs again today I'll attach gdb to it and see whether
> it's stuck on a mutex.
>
> Meetesh
>
>
> On 5/22/06, Craig A. James <cjames(at)modgraph-usa(dot)com> wrote:
> >
> > Meetesh Karia wrote:
> > > Hi all,
> > >
> > > We've recently started having a problem where a query that normally
> > > executes in ~15ms starts to take upwards of 20s to complete. When the
> > > connection that ran query is returned to the connection pool, it
> > appears
> > > as though a transaction is still in progress so the connection pool
> > > tries to cancel the transaction and close the connection. This fails
> > > and the connection is removed from the connection pool. At this
> > point,
> > > the situation rapidly degrades and we run out of connections to the
> > > postgres server.
> > >
> > > An inspection of the pg_stat_activity table shows that practically
> > every
> > > connection is running the above-mentioned query and some of those
> > > queries have been active for many minutes! We've looked at the
> > pg_locks
> > > table as well and the only exclusive locks are on transactions that
> > are
> > > open. All other locks are AccessShareLocks. Also, as far as we can
> > > tell (from looking at the Hibernate stats), every db session that is
> > > opened is closed.
> > >
> > > When this happens, if I kill one of the running postgres processes
> > (just
> > > by picking the last process returned from "ps -ef | grep postgres"),
> > the
> > > other queries will immediately finish and the system will respond.
> > > However, within 15 minutes, we'll be back in the same state as before.
> > > At that point, I've cycled Apache, Tomcat and Postgres and the system
> > > then seems to come back.
> >
> > This sounds suspiciously like a question I asked a few weeks ago, on
> > April 4. I have a process that just gets stuck. After some questions from
> > various of the experts in this forum, I used gdb(1) to attach to one of the
> > frozen Postgress backend processes, and here's what I found:
> >
> > On 5/12/2006, I wrote:
> > > Thanks, good advice. You're absolutely right, it's stuck on a
> > > mutex. After doing what you suggest, I discovered that the query
> > > in progress is a user-written function (mine). When I log in as
> > > root, and use "gdb -p <pid>" to attach to the process, here's
> > > what I find. Notice the second function in the stack, a mutex
> > > lock:
> > >
> > > (gdb) bt
> > > #0 0x0087f7a2 in _dl_sysinfo_int80 () from /lib/ld- linux.so.2
> > > #1 0x0096cbfe in __lll_mutex_lock_wait () from /lib/tls/libc.so.6
> > > #2 0x008ff67b in _L_mutex_lock_3220 () from /lib/tls/libc.so.6
> > > #3 0x4f5fc1b4 in ?? ()
> > > #4 0x00dc5e64 in std::string::_Rep::_S_empty_rep_storage () from
> > /usr/local/pgsql/lib/libchmoogle.so
> > > #5 0x009ffcf0 in ?? () from /usr/lib/libz.so.1
> > > #6 0xbfe71c04 in ?? ()
> > > #7 0xbfe71e50 in ?? ()
> > > #8 0xbfe71b78 in ?? ()
> > > #9 0x009f7019 in zcfree () from /usr/lib/libz.so.1
> > > #10 0x009f7019 in zcfree () from /usr/lib/libz.so.1
> > > #11 0x009f8b7c in inflateEnd () from /usr/lib/libz.so.1
> > > #12 0x00c670a2 in ~basic_unzip_streambuf (this=0xbfe71be0) at
> > zipstreamimpl.h:332
> > > #13 0x00c60b61 in OpenBabel::OBConversion::Read (this=0x1,
> > pOb=0xbfd923b8, pin=0xffffffea) at istream:115
> > > #14 0x00c60fd8 in OpenBabel::OBConversion::ReadString (this=0x8672b50,
> > pOb=0xbfd923b8) at obconversion.cpp:780
> > > #15 0x00c19d69 in chmoogle_ichem_mol_alloc () at stl_construct.h:120
> > > #16 0x00c1a203 in chmoogle_ichem_normalize_parent () at
> > stl_construct.h:120
> > > #17 0x00c1b172 in chmoogle_normalize_parent_sdf () at vector.tcc:243
> > > #18 0x0810ae4d in ExecMakeFunctionResult ()
> > > #19 0x0810de2e in ExecProject ()
> > > #20 0x08115972 in ExecResult ()
> > > #21 0x08109e01 in ExecProcNode ()
> > > #22 0x00000020 in ?? ()
> > > #23 0xbed4b340 in ?? ()
> > > #24 0xbf92d9a0 in ?? ()
> > > #25 0xbed4b0c0 in ?? ()
> > > #26 0x00000000 in ?? ()
> > >
> > > It looks to me like my code is trying to read the input parameter
> > > (a fairly long string, maybe 2K) from a buffer that was gzip'ed
> > > by Postgres for the trip between the client and server... somewhere
> > > along the way, a mutex gets set, and then ... it's stuck forever.
> > >
> > > ps(1) shows that this thread had been running for about 7 hours,
> > > and the job status showed that this function had been
> > > successfully called about 1 million times, before this mutex lock
> > > occurred.
> >
> > This is not an issue that's been resolved. Nobody had ever seen this
> > before. Tom Lane suggested it might be a libc/c++ bug, but unfortunately in
> > my case this lockup occurs so rarely (every few days) that it will be very
> > difficult to know if we've fixed the problem.
> >
> > If gdb(1) reveals that your process is stuck in a mutex, then you might
> > have a better chance testing this hypothesis, since your problem happens
> > within 15 minutes or so.
> >
> > Did this start recently, perhaps right after a kernel update?
> >
> > Craig
> >
>
>
From | Date | Subject | |
---|---|---|---|
Next Message | Merlin Moncure | 2006-06-21 02:56:32 | Re: Big array speed issues |
Previous Message | Markus Schaber | 2006-06-20 15:12:16 | Re: lowering priority automatically at connection |