MongoDB http://www.mondodb.org<http://www.mondodb.org/>
Michael D. Black Senior Scientist NG Information Systems Advanced Analytics Directorate ________________________________ From: sqlite-users-boun...@sqlite.org [sqlite-users-boun...@sqlite.org] on behalf of Jaco Breitenbach [jjbreitenb...@gmail.com] Sent: Tuesday, August 09, 2011 8:51 AM To: General Discussion of SQLite Database Subject: EXT :Re: [sqlite] Maximum number of tables in a data file Hi Gabriel, Is there such a database that is both free and non-GPL that you can recommend? Jaco On 9 August 2011 14:38, gabriel.b...@gmail.com <gabriel.b...@gmail.com>wrote: > Heve you ever considere using a NOSQL database I think it would serve you > better > > 2011/8/9 Jaco Breitenbach <jjbreitenb...@gmail.com> > > > Hi Igor and Michael, > > > > Yes, of course, 1440 minutes in a day. :-) > > > > I am building an application that filters out duplicate input data by > > generating an MD5 hash of each input, and implicitly comparing that > against > > a set of keys already stored in the SQLite database by doing an insert > into > > a unique-indexed table. If the insert fails, a duplicate is assumed, > > otherwise the new unique key is stored, and the input processed. > > > > The problem that I'm facing, is that I would ultimately need to process > > 1,000,000,000 records a day, with history to be kept for up to 128 days. > I > > am currently creating a new data file per day, with hourly tables. > > However, > > that will eventually result in 40,000,000+ records to be inserted into a > > single indexed table. Unfortunately the performance rate of the inserts > > into the indexed tables decreases significantly as the number of records > in > > the tables increases. This seems to be because of a CPU bottleneck > rather > > than I/O while doing the searches. > > > > I am now considering partitioning the data even further into tables that > > span shorter time periods, e.g. 60 min, 30 min, 15 min, 5 min, 1 min. I > am > > hoping that reducing the search space will help to maintain a higher > insert > > rate. > > > > I'd appreciate any feedback and comments on my suggested approach. > > > > Regards, > > Jaco > > > > > > On 9 August 2011 14:13, Igor Tandetnik <itandet...@mvps.org> wrote: > > > > > Jaco Breitenbach <jjbreitenb...@gmail.com> wrote: > > > > Can anyone please tell me if there is a limit to the number of tables > > > that > > > > can be held in a single data file? I am considering an application > > that > > > > will require a table for every minute in a day, i.e. 3600+ tables in > a > > > > single database or data file. > > > > > > First, there are 1440 minutes in a day. Second, you should be able to > > > create this number of tables: if the limit exists, it's likely much > > higher > > > than that. Finally, I predict that the schema you envision would be > very > > > awkward to work with. Have you considered a single table having > > MinuteOfDay > > > as an extra column? > > > -- > > > Igor Tandetnik > > > > > > _______________________________________________ > > > sqlite-users mailing list > > > sqlite-users@sqlite.org > > > http://sqlite.org:8080/cgi-bin/mailman/listinfo/sqlite-users > > > > > _______________________________________________ > > sqlite-users mailing list > > sqlite-users@sqlite.org > > http://sqlite.org:8080/cgi-bin/mailman/listinfo/sqlite-users > > > _______________________________________________ > sqlite-users mailing list > sqlite-users@sqlite.org > http://sqlite.org:8080/cgi-bin/mailman/listinfo/sqlite-users > _______________________________________________ sqlite-users mailing list sqlite-users@sqlite.org http://sqlite.org:8080/cgi-bin/mailman/listinfo/sqlite-users _______________________________________________ sqlite-users mailing list sqlite-users@sqlite.org http://sqlite.org:8080/cgi-bin/mailman/listinfo/sqlite-users