Jaco Breitenbach <[email protected]> wrote: > I am building an application that filters out duplicate input data by > generating an MD5 hash of each input, and implicitly comparing that against > a set of keys already stored in the SQLite database by doing an insert into > a unique-indexed table. If the insert fails, a duplicate is assumed, > otherwise the new unique key is stored, and the input processed. > > The problem that I'm facing, is that I would ultimately need to process > 1,000,000,000 records a day, with history to be kept for up to 128 days. I > am currently creating a new data file per day, with hourly tables.
Doesn't that defeat the point? Wouldn't that only guarantee uniqueness within the last hour? -- Igor Tandetnik _______________________________________________ sqlite-users mailing list [email protected] http://sqlite.org:8080/cgi-bin/mailman/listinfo/sqlite-users

