Hi Alexander, I looked into my gam-stderr.log file and found *Error: Could not find or load main class org.apache.giraph.yarn.GiraphApplicationMaster* As you have already mentioned in your reply about that for loop in *hadoop-env.sh. * So can you please tell me what is *mylib *and *share/myLib/giraph** ? *Because I am having only doc and hadoop directories inside $HADOOP_HOME/share/
Also how to get *giraph-1.1.0-hadoop-2.4.0.jar *and* giraph-examples-1.1.0-hadoop-*2.4.0.jar (without the dependencies) as I am having only giraph-examples-1.1.0-SNAPSHOT-for-hadoop-2.4.0-jar-with-dependencies.jar inside giraph-examples/target/munged/. Regards, Vikalp Handa On Tue, Aug 12, 2014 at 4:25 PM, Alexander Sirotin <sirot...@web.de> wrote: > Hello Vikalp, > > you are showing the output of the client. The errors I ve showed were > coming from the ResourceManager, the GiraphApplicationMaster > (gam-stderr.log) and the GiraphYarnTask (TaskRunner). Checkout these > logfiles on your systems. > > In case you get an error-message "class GiraphApplicationMaster not > found": I solved it by adding a for-loop in hadoop-env.sh: > > for f in `find $HADOOP_HOME/share/myLib/ -name \*.jar`; do > if [ "$HADOOP_CLASSPATH" ]; then > export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f > else > export HADOOP_CLASSPATH=$f > fi > done > > In share/myLib/giraph I have put giraph-1.1.0-hadoop-2.4.0.jar and > giraph-examples-1.1.0-hadoop-2.4.0.jar (without the dependencies). > Additionally I copied every jar from /usr/local/giraph/lib also to this > folder. > > Then you can restart the cluster and every instance should know giraph. > > After this it should function, if you have enough memory. I get then a > message, that the GiraphApplicationMaster dont have enough memory and then > get killed. But before this, it has started the TaskRunner, which keeps the > whole job endlessly alive. > > Sincerely, > > > > > On 12.08.2014 11:55, Vikalp Handa wrote: > > @Alexander Sirotin : Thanks for your reply. I am really sorry I haven't > faced this problem after I executed it. Rather I am now having a different > issue with Containers : > > *Result : * > 14/08/12 15:17:51 INFO yarn.GiraphYarnClient: > ApplicationSumbissionContext for GiraphApplicationMaster launch container > is populated. > 14/08/12 15:17:51 INFO yarn.GiraphYarnClient: Submitting application to ASM > 14/08/12 15:17:52 INFO impl.YarnClientImpl: Submitted application > application_1407836750214_0001 > 14/08/12 15:17:52 INFO yarn.GiraphYarnClient: Got new appId after > submission :application_1407836750214_0001 > 14/08/12 15:17:52 INFO yarn.GiraphYarnClient: GiraphApplicationMaster > container request was submitted to ResourceManager for job: Giraph: > org.apache.giraph.examples.SimpleShortestPathsComputation > 14/08/12 15:17:52 INFO yarn.GiraphYarnClient: Giraph: > org.apache.giraph.examples.SimpleShortestPathsComputation, Elapsed: 0.99 > secs > 14/08/12 15:17:52 INFO yarn.GiraphYarnClient: > appattempt_1407836750214_0001_000001, State: ACCEPTED, Containers used: 1 > 14/08/12 15:17:56 INFO yarn.GiraphYarnClient: Giraph: > org.apache.giraph.examples.SimpleShortestPathsComputation, Elapsed: 5.01 > secs > 14/08/12 15:17:56 INFO yarn.GiraphYarnClient: > appattempt_1407836750214_0001_000002, State: ACCEPTED, Containers used: 0 > 14/08/12 15:18:00* ERROR yarn.GiraphYarnClient: Giraph: > org.apache.giraph.examples.SimpleShortestPathsComputation reports FAILED > state, diagnostics show: Application application_1407836750214_0001 failed > 2 times due to AM Container for appattempt_1407836750214_0001_000002 exited > with exitCode: 1 due to: Exception from container-launch: > org.apache.hadoop.util.Shell$ExitCodeException: * > *org.apache.hadoop.util.Shell$ExitCodeException:* > at org.apache.hadoop.util.Shell.runCommand(Shell.java:505) > at org.apache.hadoop.util.Shell.run(Shell.java:418) > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650) > at > org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195) > at > org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300) > at > org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > > > Container exited with a non-zero exit code 1 > .Failing this attempt.. Failing the application. > 14/08/12 15:18:00 INFO yarn.GiraphYarnClient: Cleaning up HDFS distributed > cache directory for Giraph job. > 14/08/12 15:18:00 INFO yarn.GiraphYarnClient: Completed Giraph: > org.apache.giraph.examples.SimpleShortestPathsComputation: FAILED, total > running time: 0 minutes, 7 seconds. > > I have also checked my *yarn-site.xml* file and updated with the > following property - value pairs inside configuration: > > <property> > <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> > <value>org.apache.hadoop.mapred.ShuffleHandler</value> > </property> > <property> > <name>yarn.application.classpath</name> > <value> > %HADOOP_HOME%\etc\hadoop, > %HADOOP_HOME%\share\hadoop\common\*, > %HADOOP_HOME%\share\hadoop\common\lib\*, > %HADOOP_HOME%\share\hadoop\hdfs\*, > %HADOOP_HOME%\share\hadoop\hdfs\lib\*, > %HADOOP_HOME%\share\hadoop\mapreduce\*, > %HADOOP_HOME%\share\hadoop\mapreduce\lib\*, > %HADOOP_HOME%\share\hadoop\yarn\*, > %HADOOP_HOME%\share\hadoop\yarn\lib\* > </value> > </property> > > > > Regards, > Vikalp Handa > > >