From: | Tom Lane <tgl(at)sss(dot)pgh(dot)pa(dot)us> |
---|---|
To: | edipoelder(at)ig(dot)com(dot)br |
Cc: | pgsql-sql(at)postgresql(dot)org |
Subject: | Re: Memory and performance |
Date: | 2001-04-05 01:53:30 |
Message-ID: | 14398.986435610@sss.pgh.pa.us |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-sql |
edipoelder(at)ig(dot)com(dot)br writes:
> I have noted that Postgresql don't make a good memory handle. I have
> made the tables/procedure (in attached file) and run it as "select bench(10,
> 5000)". This will give a 50000 records inserts (5 x 10000). (well, I run it
> on a P200+64MB of RAM, under Linux, and Postgres 7.0.2. In a more powerfull
> machine, you can try other values).
> I get as result, the following times:
> id | objname | benchtime
> ----+---------+-----------
> 1 | group 1 | 00:00:32
> 2 | group 2 | 00:00:47
> 3 | group 3 | 00:01:13
> 4 | group 4 | 00:01:41
> 5 | group 5 | 00:02:08
> (5 rows)
This is an inefficiency in handling of foreign-key triggers. It's fixed
for 7.1 --- in current sources I get
id | objname | benchtime
----+----------+-----------
1 | group 1 | 00:00:03
2 | group 2 | 00:00:03
3 | group 3 | 00:00:03
4 | group 4 | 00:00:03
5 | group 5 | 00:00:03
6 | group 6 | 00:00:03
7 | group 7 | 00:00:03
8 | group 8 | 00:00:03
9 | group 9 | 00:00:03
10 | group 10 | 00:00:03
(10 rows)
regards, tom lane
From | Date | Subject | |
---|---|---|---|
Next Message | Joel Burton | 2001-04-05 01:59:51 | [Tip] Using list-aggregates for faster batching |
Previous Message | Gerald Gutierrez | 2001-04-05 00:24:12 | RE: RE: serial type; race conditions |