Re: how to make duplicate finding query faster?

From: Holger Jakobs <holger(at)jakobs(dot)com>
To: pgsql-admin(at)lists(dot)postgresql(dot)org
Subject: Re: how to make duplicate finding query faster?
Date: 2020-12-30 08:23:42
Message-ID: 5bf642e4-eb22-8eba-e96d-5b33b5010b03@jakobs.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-admin

Am 30.12.20 um 08:36 schrieb Sachin Kumar:
> Hi All,
>
> I am uploading data into PostgreSQL using the CSV file and checking if
> there is any duplicates value in DB it should return a duplicate
> error.  I am using below mention query.
>
> if Card_Bank.objects.filter( Q(ACCOUNT_NUMBER=card_number) ).exists():
>         flag=2
>       else:
>         flag=1
> it is taking too much time i am using 600k cards in CSV.
>
> Kindly help me in making the query faster.
>
> I am using Python, Django & PostgreSQL.
> --
>
> Best Regards,
> Sachin Kumar

I think it would be easier to not check the duplicates before, but let
the DB complain about duplicates.

That would about slash the roundtrips to the DB in half. Instead of
check + insert there would be only an insert, which might fail every now
and then.

Regards,

Holger

--
Holger Jakobs, Bergisch Gladbach, Tel. +49-178-9759012

In response to

Responses

Browse pgsql-admin by date

  From Date Subject
Next Message Sachin Kumar 2020-12-30 08:59:08 Re: how to make duplicate finding query faster?
Previous Message Sachin Kumar 2020-12-30 07:36:40 how to make duplicate finding query faster?