On Fri, Mar 29, 2013 at 12:17 PM, Jeff Archer <jsarc...@nanotronicsimaging.com> wrote: > I have previously made an apparently bad assumption about this so now I > would like to go back to the beginning of the problem and ask the most > basic question first without any preconceived ideas. > > This use case is from an image processing application. I have a large > amount of intermediate data (way exceeds physical memory on my 24GB > machine). So, I need to store it temporarily on disk until getting to next > phase of processing. I am planning to use a large SSD dedicated to holding > this temporary data. I do not need any recoverability in case of hardware, > power or other failure. Each item to be stored is 9 DWORDs, 4 doubles and > 2 variable sized BLOBS which are images. > > I could write directly to a file myself. But I would need to provide some > minimal indexing, some amount of housekeeping to manage variable > sized BLOBS and some minimal synchronization so that multiple instances of > the same application could operate simultaneously on a single set of data. > > So, then I though that SQLite could manage these things nicely for me so > that I don't have to write and debug indexing and housekeeping code that > already exists in SQLite. > > So, question is: What is the way to get the fastest possible performance > from SQLite when I am willing to give up all recoverability guarantees?
Use pragma journal_mode = off; pragma synchronous = off; pragma locking_mode = exclusive; In addition to that you may issue BEGIN statement at the beginning of the application and never COMMIT. Pavel _______________________________________________ sqlite-users mailing list sqlite-users@sqlite.org http://sqlite.org:8080/cgi-bin/mailman/listinfo/sqlite-users