For some reason I don't see that reply from Jonathan in my Inbox. I'll try to google it.
What should be my next step in that case? I can't use pig then? On Thu, May 26, 2011 at 10:00 AM, Harsh J <ha...@cloudera.com> wrote: > I think Jonathan Coveney's reply on user@pig answered your question. > Its basically an issue of hadoop version differences between the one > Pig 0.8.1 release got bundled with vs. Hadoop 0.20.203 release which > is newer. > > On Thu, May 26, 2011 at 10:26 PM, Mohit Anchlia <mohitanch...@gmail.com> > wrote: >> I sent this to pig apache user mailing list but have got no response. >> Not sure if that list is still active. >> >> thought I will post here if someone is able to help me. >> >> I am in process of installing and learning pig. I have a hadoop >> cluster and when I try to run pig in mapreduce mode it errors out: >> >> Hadoop version is hadoop-0.20.203.0 and pig version is pig-0.8.1 >> >> Error before Pig is launched >> ---------------------------- >> ERROR 2999: Unexpected internal error. Failed to create DataStorage >> >> java.lang.RuntimeException: Failed to create DataStorage >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:75) >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.<init>(HDataStorage.java:58) >> at >> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:214) >> at >> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:134) >> at org.apache.pig.impl.PigContext.connect(PigContext.java:183) >> at org.apache.pig.PigServer.<init>(PigServer.java:226) >> at org.apache.pig.PigServer.<init>(PigServer.java:215) >> at org.apache.pig.tools.grunt.Grunt.<init>(Grunt.java:55) >> at org.apache.pig.Main.run(Main.java:452) >> at org.apache.pig.Main.main(Main.java:107) >> Caused by: java.io.IOException: Call to dsdb1/172.18.60.96:54310 >> failed on local exception: java.io.EOFException >> at org.apache.hadoop.ipc.Client.wrapException(Client.java:775) >> at org.apache.hadoop.ipc.Client.call(Client.java:743) >> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >> at $Proxy0.getProtocolVersion(Unknown Source) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) >> at >> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106) >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207) >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82) >> at >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378) >> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390) >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196) >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95) >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:72) >> ... 9 more >> Caused by: java.io.EOFException >> at java.io.DataInputStream.readInt(DataInputStream.java:375) >> at >> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501) >> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446) >> > > > > -- > Harsh J >