Hi all, I am running the PageRank code on single machine hadoop installation. In particular, I am running the code using four workers on a graph of 100.000 nodes. I am getting this exception:
012-09-13 22:54:38,379 INFO org.apache.hadoop.mapred.Task: Communication exception: java.lang.OutOfMemoryError: GC overhead limit exceeded at org.apache.hadoop.util.ProcfsBasedProcessTree.getProcessList(ProcfsBasedProcessTree.java:365) at org.apache.hadoop.util.ProcfsBasedProcessTree.getProcessTree(ProcfsBasedProcessTree.java:138) at org.apache.hadoop.util.LinuxResourceCalculatorPlugin.getProcResourceValues(LinuxResourceCalculatorPlugin.java:401) at org.apache.hadoop.mapred.Task.updateResourceCounters(Task.java:808) at org.apache.hadoop.mapred.Task.updateCounters(Task.java:830) at org.apache.hadoop.mapred.Task.access$600(Task.java:66) at org.apache.hadoop.mapred.Task$TaskReporter.run(Task.java:666) at java.lang.Thread.run(Thread.java:662) It's seems to be a problem related to communication stuff. Do I have to specify or modify some options to hadoop to avoid this? Best, Franco Maria