> On 1 Jul 2014, at 12:26pm, Gert Van Assche <ger...@gmail.com> wrote: > > 1 - Open the BIG db, attach all small files, ten by 10, and copy the tables > from the attached databases to the big table. I can speed up the import by > putting the INSERT in a transaction. > > 2 - Export one by one the table from all databases to a CSV file; import > all CSV files into the BIG db; I don't have to worry about grouping the > files by 10, but I cannot use transaction. > > What would be the best way to do this? One of these 2 ways, or is there a > third way to do this?
First, are you doing this just once, or does it have to become part of a workflow for regular use ? The results of the two should be exactly equivalent. (1) will be faster. You can do things like "INSERT INTO tableB SELECT * FROM tableA". This automatically executes the whole thing as one transaction. (2) will be easier to debug. SQLite doesn't have a built-in function to express a row as an INSERT command, but this is part of the function of the SQLite shell tool and shell tool scripts can incorporate "BEGIN" and "END" commands. If I was doing it just once I'd probably write a script for the shell tool that did (2). If I had to implement it as part of a frequently used workflow I'd probably write code which did (1). Simon. _______________________________________________ sqlite-users mailing list sqlite-users@sqlite.org http://sqlite.org:8080/cgi-bin/mailman/listinfo/sqlite-users