From: | Stephan Szabo <sszabo(at)megazone23(dot)bigpanda(dot)com> |
---|---|
To: | Hervé Piedvache <herve(at)elma(dot)fr> |
Cc: | "pgsql-general(at)postgresql(dot)org" <pgsql-general(at)postgresql(dot)org> |
Subject: | Re: Vacuum analyze with 7 millions records ... takes 1 |
Date: | 2001-10-12 19:09:02 |
Message-ID: | Pine.BSF.4.21.0110121207530.97792-100000@megazone23.bigpanda.com |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-general |
On Fri, 12 Oct 2001, [iso-8859-1] Herv Piedvache wrote:
> I have a table (not so big) 7 millions records ... about 13 fields
> inside and 5 indexes ...
> I'm usign a Bi-Pentium III 800 Mhz with 1Gb under Linux kernel 2.4.10
> ... and the database do nothing else !
>
> Inside this table I insert about 10000 records every days, and have to
> update some records (about 3 to 10 000) by day.
>
> So if I want a good timing access to my database I have to vacuum
> analyze it every day ...
>
> But the vacuum analyze take about 1 hour ... exactly 59 minutes ... ;) I
> hour the table is locked ... so impossible to be used :(
>
> Do you have any idea to get a better result for the vacuum ... ? And why
> it take 1 hour ?
>
> Or any solutionq to get good timing and better results ? ;)
Hmm, I'm not sure if it'll help, but have you tried dropping the indexes,
doing the vacuum and recreating them? If nothing else, it might cut the
unavailable time down.
From | Date | Subject | |
---|---|---|---|
Next Message | Charles Tassell | 2001-10-12 19:10:49 | Re: COMMIT after an ERROR? |
Previous Message | David Link | 2001-10-12 19:03:08 | Re: Understanding explain costs |