Argh, I'll look into it and see where Grams are kept in memory. There really shouldn't be any place where they're retained beyond what's needed for a single document. I doubt that there are documents in wikipedia that would blow the heap in this way, but I supposed it's possible. You're just doing bigrams, or did you end up going up to 5-grams?
On Sun, Feb 28, 2010 at 7:50 AM, Robin Anil <robin.a...@gmail.com> wrote: > after 9 hours of compute, it failed. It never went past the colloc combiner > pass :( > > reason. I will have to tag drew along to identify the possible cause of this > out of memory error > > > java.lang.OutOfMemoryError: Java heap space > at > org.apache.mahout.utils.nlp.collocations.llr.Gram.<init>(Gram.java:67) > at > org.apache.mahout.utils.nlp.collocations.llr.CollocCombiner.reduce(CollocCombiner.java:62) > at > org.apache.mahout.utils.nlp.collocations.llr.CollocCombiner.reduce(CollocCombiner.java:30) > at > org.apache.hadoop.mapred.MapTask$MapOutputBuffer.combineAndSpill(MapTask.java:921) > at > org.apache.hadoop.mapred.MapTask$MapOutputBuffer.mergeParts(MapTask.java:1077) > at > org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:719) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:233) > at > org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2216) >