like i said only 5 mil articles. Maybe you can generate a co-occurrence matrix :) every ngram to every other ngram :) Sounds fun? It will be HUGE!
On Sun, Feb 28, 2010 at 3:43 AM, Jake Mannix <jake.man...@gmail.com> wrote: > 15GB of tokenized documents, not bad, not bad. We're not going > to get a multi-billion entry matrix out of this though, are we? > > -jake > > On Sat, Feb 27, 2010 at 2:06 PM, Robin Anil <robin.a...@gmail.com> wrote: > > > Update: > > > > in 20 mins the tokenization stage is complete But its not evident in the > > online UI. > > I found it by checking the s3 output folder > > > > 2010-02-27 21:50 2696826329 > > s3://robinanil/wikipedia/tokenized-documents/part-00000 > > 2010-02-27 21:52 2385184391 > > s3://robinanil/wikipedia/tokenized-documents/part-00001 > > 2010-02-27 21:52 2458566158 > > s3://robinanil/wikipedia/tokenized-documents/part-00002 > > 2010-02-27 21:53 2500213973 > > s3://robinanil/wikipedia/tokenized-documents/part-00003 > > 2010-02-27 21:50 2533593862 > > s3://robinanil/wikipedia/tokenized-documents/part-00004 > > 2010-02-27 21:54 3580695441 > > s3://robinanil/wikipedia/tokenized-documents/part-00005 > > 2010-02-27 22:02 0 > > s3://robinanil/wikipedia/tokenized-documents_$folder$ > > 2010-02-27 22:02 0 > > s3://robinanil/wikipedia/wordcount/subgrams/_temporary_$folder$ > > 2010-02-27 22:02 0 > > s3://robinanil/wikipedia/wordcount/subgrams_$folder$ > > 2010-02-27 22:02 0 s3://robinanil/wikipedia/wordcount_$folder$ > > >