that because SQLLite isn't a multi-user DB server but a single user embedded (desktop) db. Use the right tool for the job.
On Saturday, October 15, 2016 at 7:02:58 PM UTC+2, Ralph Smith wrote: > > How are the processes supposed to interact with the database? Without > extra synchronization logic, SQLite.jl gives (occasionally) > ERROR: LoadError: On worker 2: > SQLite.SQLiteException("database is locked") > which on the face of it suggests that all workers are using the same > connection, although I opened the DB separately in each process. > (I think we should get "busy" instead of "locked", but then still have no > good way to test for this and wait for a wake-up signal.) > So we seem to be at least as badly off as the original post, except with > DB calls instead of simple writes. > > We shouldn't have to stand up a separate multithreaded DB server just for > this. Would you be kind enough to give us an example of simple (i.e. not > client-server) multiprocess DB access in Julia? > > On Saturday, October 15, 2016 at 9:40:17 AM UTC-4, Steven Sagaert wrote: >> >> It still surprises me how in the scientific computing field people still >> refuse to learn about databases and then replicate database functionality >> in files in a complicated and probably buggy way. HDF5 is one example, >> there are many others. If you want to to fancy search (i.e. speedup search >> via indices) or do things like parallel writes/concurrency you REALLY >> should use databases. That's what they were invented for decades ago. >> Nowadays there a bigger choice than ever: Relational or non-relational >> (NOSQL), single host or distributed, web interface or not, disk-based or >> in-memory,... There really is no excuse anymore not to use a database if >> you want to go beyond just reading in a bunch of data in one go in memory. >> >> On Monday, October 10, 2016 at 5:09:39 PM UTC+2, Zachary Roth wrote: >>> >>> Hi, everyone, >>> >>> I'm trying to save to a single file from multiple worker processes, but >>> don't know of a nice way to coordinate this. When I don't coordinate, >>> saving works fine much of the time. But I sometimes get errors with >>> reading/writing of files, which I'm assuming is happening because multiple >>> processes are trying to use the same file simultaneously. >>> >>> I tried to coordinate this with a queue/channel of `Condition`s managed >>> by a task running in process 1, but this isn't working for me. I've tried >>> to simiplify this to track down the problem. At least part of the issue >>> seems to be writing to the channel from process 2. Specifically, when I >>> `put!` something onto a channel (or `push!` onto an array) from process 2, >>> the channel/array is still empty back on process 1. I feel like I'm >>> missing something simple. Is there an easier way to go about coordinating >>> multiple processes that are trying to access the same file? If not, does >>> anyone have any tips? >>> >>> Thanks for any help you can offer. >>> >>> Cheers, >>> ---Zachary >>> >>