100GB of RAM is practically common. Recently I’ve seen many indicators and item metadata stored with cooccurrence and indexed. This produces extremely flexible results since the query determines the result, not the model. But it does increase the number of cooccurrences linearly with # of indicator types.
As to DB, any suggestions? It would need to have a very high performance memory cached implementation. I wonder if the search engine itself would work. This would at least reduce the number of subsystems to deal with. On Apr 24, 2015, at 4:13 PM, Ted Dunning <ted.dunn...@gmail.com> wrote: Sounds about right. My guess is that memory is now large enough, especially on a cluster that the cooccurrence will fit into memory quite often. Taking a large example of 10 million items and 10,000 cooccurrences each, there will be 100 billion cooccurrences to store which shouldn't take more than about half a TB of data if fully populated. This isn't that outrageous any more. With SSD's as backing store, even 100GB of RAM or less might well produce very nice results. Depending on incoming transaction rates, using spinning disk as a backing store might also work with small memory. Experiments are in order. On Fri, Apr 24, 2015 at 8:12 AM, Pat Ferrel <p...@occamsmachete.com> wrote: > Ok, seems right. > > So now to data structures. The input frequency vectors need to be paired > with each input interaction type and would be nice to have as something > that can be copied very fast as they get updated. Random access would also > be nice but iteration is not needed. Over time they will get larger as all > items get interactions, users will get more actions and appear in more > vectors (with multi-intereaction data). Seems like hashmaps? > > The cooccurrence matrix is more of a question to me. It needs to be > updatable at the row and column level, and random access for both row and > column would be nice. It needs to be expandable. To keep it small the keys > should be integers, not full blown ID strings. There will have to be one > matrix per interaction type. It should be simple to update the Search > Engine to either mirror the matrix of use it directly for index updates. > Each indicator update should cause an index update. > > Putting aside speed and size issues this sounds like a NoSQL DB table that > is cached in-memeory. > > On Apr 23, 2015, at 3:04 PM, Ted Dunning <ted.dunn...@gmail.com> wrote: > > On Thu, Apr 23, 2015 at 8:53 AM, Pat Ferrel <p...@occamsmachete.com> wrote: > >> This seems to violate the random choice of interactions to cut but now >> that I think about it does a random choice really matter? >> > > It hasn't ever mattered such that I could see. There is also some reason > to claim that earliest is best if items are very focussed in time. Of > course, the opposite argument also applies. That leaves us with empiricism > where the results are not definitive. > > So I don't think that it matters, but I don't think that it does. > >