Re: how to make duplicate finding query faster?

Поиск
Список
Период
Сортировка
От Scott Ribe
Тема Re: how to make duplicate finding query faster?
Дата
Msg-id FEEE2DC4-B506-4B34-80FE-07FCC0ADC61E@elevated-dev.com
обсуждение исходный текст
Ответ на how to make duplicate finding query faster?  (Sachin Kumar <sachinkumaras@gmail.com>)
Ответы Re: how to make duplicate finding query faster?
Список pgsql-admin
> On Dec 30, 2020, at 12:36 AM, Sachin Kumar <sachinkumaras@gmail.com> wrote:
>
> Hi All,
>
> I am uploading data into PostgreSQL using the CSV file and checking if there is any duplicates value in DB it should
returna duplicate error.  I am using below mention query. 
>
> if Card_Bank.objects.filter( Q(ACCOUNT_NUMBER=card_number) ).exists():
>         flag=2
>       else:
>         flag=1
> it is taking too much time i am using 600k cards in CSV.
>
> Kindly help me in making the query faster.
>
> I am using Python, Django & PostgreSQL.
> --
>
> Best Regards,
> Sachin Kumar

Are you checking one-by-one because your goal is not to fail the whole upload that contains the duplicates, but rather
toskip only the duplicates? 

If that's the case, I think you'd be better off copying the CSV straight into a temp table, using a join to delete
duplicatesfrom it, then insert the remainder into the target table, and finally drop the temp table. 


В списке pgsql-admin по дате отправления:

Предыдущее
От: "Gavan Schneider"
Дата:
Сообщение: Re: how to make duplicate finding query faster?
Следующее
От: Sachin Kumar
Дата:
Сообщение: Re: how to make duplicate finding query faster?