Does anyone know the reason? Thank you very much!!
On Mon, Jul 28, 2014 at 1:29 PM, Jing Fan <fanjin...@gmail.com> wrote: > Hi everyone, > > I was running a job with outOfCore option and Giraph get stuck. > > The RAM is 96G and the dataset is 20G. The server has 16 physical cores. > I was running PageRank with the following command > > /usr/local/hadoop/bin/hadoop jar > /usr/local/giraph/giraph-examples/target/giraph-examples-1.1.0-SNAPSHOT-for-hadoop-1.2.1-jar-with-dependencies.jar > org.apache.giraph.GiraphRunner -D giraph.useOutOfCoreGraph=true -D > giraph.useOutofCoreMessages=true -D giraph.maxPartitionsInMemory=50 -D > giraph.maxMessagesInMemory=500000000 -Dgiraph.zkMinSessionTimeout=9000000 > -Dgiraph.zkMaxSessionTimeout=9000000 -Dgiraph.zkSessionMsecTimeout=9000000 > org.apache.examples.PageRank -eif > org.apache.giraph.io.formats.IntNullTextEdgeInputFormat -eip > /user/hduser/xx.txt -w 14 > > It get stuck with: > > 14/07/28 22:05:58 INFO job.JobProgressTracker: Data from 14 workers - > Compute superstep 1: 38731088 out of 41652230 vertices computed; 182 out of > 196 partitions computed; min free memory on worker 4 - 768.51MB, average > 1296.96MB > and then was killed. > > I check the mem, it uses nearly 100% of the memory. > > Does anyone know the reason? > > Thanks! >