am  25.05.2005, um 13:58:07 -0300 mailte [EMAIL PROTECTED] folgendes:
> Hi.
> Thanks for the article...
> But, I have read it and the query works very slow...
> My table have aprox. 180.000 records (correct) and in entire table it has
> aprox.360.000 records(duplicated)...

How often is this necessary?


> Is there a way to delete those duplicated records faster??? Remembering the
> table have aprox 360.000 records...

I dont know, but i think, you should prevent duplicated records in the
future, and make the job (delete duplicates) now.

Btw.: you wrote, there is a primary key on the first row. Real?

,----[  sorry, messages in german language  ]
| test_db=# create table blub (id int primary key, name varchar);
| HINWEIS:  CREATE TABLE / PRIMARY KEY erstellt implizit einen Index 
>>blub_pkey<< für Tabelle >>blub<<
| CREATE TABLE
| test_db=# insert into blub values (1, 'x');
| INSERT 970706 1
| test_db=# insert into blub values (1, 'y');
| FEHLER:  duplizierter Schlüssel verletzt Unique-Constraint >>blub_pkey<<
`----

In other words: if there a primary key on the first row, you cannot
insert duplicates.


Regards, Andreas
-- 
Andreas Kretschmer    (Kontakt: siehe Header)
Heynitz:  035242/47212,      D1: 0160/7141639
GnuPG-ID 0x3FFF606C http://wwwkeys.de.pgp.net
 ===    Schollglas Unternehmensgruppe    === 

---------------------------(end of broadcast)---------------------------
TIP 5: Have you checked our extensive FAQ?

               http://www.postgresql.org/docs/faq

Reply via email to