sorry for wide distribution, I further check the folder '_bsp/_defaultZkManagerDir/job_1344903945125_0032' exists, and it has one sub folder '_bsp/_defaultZkManagerDir/job_1344903945125_0032/_task' and another file inside, so the hdfs file permission should not be a issue. but not sure why Giraph still complain '_bsp/_defaultZkManagerDir/job_1344903945125_0032/_zkServer does not exist'.
Does Zookeeper needs further configuration? Or any other possible reason cannot create _zkServer folder ? Thanks, Johnny On Mon, Aug 20, 2012 at 11:59 AM, Johnny Zhang <xiao...@cloudera.com> wrote: > Alessandro: > Thanks for reminding me on that. Now I can run the pagerank > example successfully, though I still get one zookeeper server related > exception. Here is part of the log: > > 12/08/20 11:56:44 WARN mapreduce.Job: Error reading task output Server > returned HTTP response code: 400 for URL: > http://cs-10-20-76-76.cloud.cloudera.com:8080/tasklog?plaintext=true&attemptid=attempt_1344903945125_0032_m_000002_2&filter=stdout > 12/08/20 11:56:44 WARN mapreduce.Job: Error reading task output Server > returned HTTP response code: 400 for URL: > http://cs-10-20-76-76.cloud.cloudera.com:8080/tasklog?plaintext=true&attemptid=attempt_1344903945125_0032_m_000002_2&filter=stderr > 12/08/20 11:56:44 INFO mapreduce.Job: Task Id : > attempt_1344903945125_0032_m_000001_2, Status : FAILED > Error: java.lang.RuntimeException: java.io.FileNotFoundException: File > _bsp/_defaultZkManagerDir/job_1344903945125_0032/_zkServer does not exist. > at > org.apache.giraph.zk.ZooKeeperManager.onlineZooKeeperServers(ZooKeeperManager.java:749) > at org.apache.giraph.graph.GraphMapper.setup(GraphMapper.java:320) > at org.apache.giraph.graph.GraphMapper.run(GraphMapper.java:570) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:725) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:152) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:147) > Caused by: java.io.FileNotFoundException: File > _bsp/_defaultZkManagerDir/job_1344903945125_0032/_zkServer does not exist. > at > org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:365) > at > org.apache.giraph.zk.ZooKeeperManager.onlineZooKeeperServers(ZooKeeperManager.java:708) > ... 9 more > > 12/08/20 11:56:44 WARN mapreduce.Job: Error reading task output Server > returned HTTP response code: 400 for URL: > http://cs-10-20-76-76.cloud.cloudera.com:8080/tasklog?plaintext=true&attemptid=attempt_1344903945125_0032_m_000001_2&filter=stdout > 12/08/20 11:56:44 WARN mapreduce.Job: Error reading task output Server > returned HTTP response code: 400 for URL: > http://cs-10-20-76-76.cloud.cloudera.com:8080/tasklog?plaintext=true&attemptid=attempt_1344903945125_0032_m_000001_2&filter=stderr > 12/08/20 11:56:45 INFO mapreduce.Job: Job job_1344903945125_0032 failed > with state FAILED due to: > 12/08/20 11:56:45 INFO mapreduce.Job: Counters: 28 > File System Counters > FILE: Number of bytes read=120 > FILE: Number of bytes written=49450 > FILE: Number of read operations=0 > FILE: Number of large read operations=0 > FILE: Number of write operations=0 > HDFS: Number of bytes read=44 > HDFS: Number of bytes written=0 > HDFS: Number of read operations=8 > HDFS: Number of large read operations=0 > HDFS: Number of write operations=3 > Job Counters > Failed map tasks=10 > Launched map tasks=13 > Other local map tasks=13 > Total time spent by all maps in occupied slots (ms)=692328 > Total time spent by all reduces in occupied slots (ms)=0 > Map-Reduce Framework > Map input records=0 > Map output records=0 > Input split bytes=44 > Spilled Records=0 > Failed Shuffles=0 > Merged Map outputs=0 > GC time elapsed (ms)=34 > CPU time spent (ms)=450 > Physical memory (bytes) snapshot=96169984 > Virtual memory (bytes) snapshot=1599012864 > Total committed heap usage (bytes)=76087296 > File Input Format Counters > Bytes Read=0 > File Output Format Counters > Bytes Written=0 > > > Thanks, > Johnny > > On Mon, Aug 20, 2012 at 11:47 AM, Alessandro Presta <alessan...@fb.com>wrote: > >> Looks like you compiled for hadoop 0.20.203, which had a different API >> (that's why we have to use Munge). Can you try recompiling with the >> hadoop_2.0.0 profile? >> >> From: Johnny Zhang <xiao...@cloudera.com> >> Reply-To: "user@giraph.apache.org" <user@giraph.apache.org> >> Date: Monday, August 20, 2012 7:31 PM >> To: "user@giraph.apache.org" <user@giraph.apache.org> >> Subject: cannot run Giraph trunk with Hadoop 2.0.0-alpha >> >> Hi, all: >> I am trying to run Giraph trunk with Hadoop 2.0.0-alpha. >> I am getting below error when I run a page rank example job with 3 >> workers. >> >> # hadoop jar >> target/giraph-0.2-SNAPSHOT-for-hadoop-0.20.203.0-jar-with-dependencies.jar >> org.apache.giraph.benchmark.PageRankBenchmark -e 1 -s 3 -v -V 50000000 -w 3 >> 12/08/20 11:10:38 WARN mapred.JobConf: The variable mapred.child.ulimit >> is no longer used. >> 12/08/20 11:10:38 INFO benchmark.PageRankBenchmark: Using class >> org.apache.giraph.benchmark.PageRankBenchmark >> 12/08/20 11:10:38 WARN conf.Configuration: mapred.job.tracker is >> deprecated. Instead, use mapreduce.jobtracker.address >> 12/08/20 11:10:38 WARN conf.Configuration: mapred.job.map.memory.mb is >> deprecated. Instead, use mapreduce.map.memory.mb >> 12/08/20 11:10:38 WARN conf.Configuration: mapred.job.reduce.memory.mb is >> deprecated. Instead, use mapreduce.reduce.memory.mb >> 12/08/20 11:10:38 WARN conf.Configuration: >> mapred.map.tasks.speculative.execution is deprecated. Instead, use >> mapreduce.map.speculative >> Exception in thread "main" java.lang.IncompatibleClassChangeError: Found >> interface org.apache.hadoop.mapreduce.JobContext, but class was expected >> at >> org.apache.giraph.bsp.BspOutputFormat.checkOutputSpecs(BspOutputFormat.java:43) >> at >> org.apache.hadoop.mapreduce.JobSubmitter.checkSpecs(JobSubmitter.java:411) >> at >> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:326) >> at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1226) >> at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1223) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:396) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232) >> at org.apache.hadoop.mapreduce.Job.submit(Job.java:1223) >> at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1244) >> at org.apache.giraph.graph.GiraphJob.run(GiraphJob.java:714) >> at >> org.apache.giraph.benchmark.PageRankBenchmark.run(PageRankBenchmark.java:150) >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >> at >> org.apache.giraph.benchmark.PageRankBenchmark.main(PageRankBenchmark.java:164) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) >> at java.lang.reflect.Method.invoke(Method.java:597) >> at org.apache.hadoop.util.RunJar.main(RunJar.java:208) >> >> >> My $HADOOP_MAPRED_HOME and $JAVA_HOME is set up correctly, could anyone >> tell me if I need to setup anything else? Thanks a lot. >> >> Johnny >> > >