On Thu, May 26, 2011 at 10:06 AM, Jonathan Coveney <jcove...@gmail.com> wrote: > I'll repost it here then :) > > "Here is what I had to do to get pig running with a different version of > Hadoop (in my case, the cloudera build but I'd try this as well):
> > build pig-withouthadoop.jar by running "ant jar-withouthadoop". Then, when > you run pig, put the pig-withouthadoop.jar on your classpath as well as your > hadoop jar. In my case, I found that scripts only worked if I additionally > manually registered the antlr jar: Thanks Jonathan! I will give it a shot. > > register /path/to/pig/build/ivy/lib/Pig/antlr-runtime-3.2.jar;" Is this a windows command? Sorry, have not used this before. > > 2011/5/26 Mohit Anchlia <mohitanch...@gmail.com> > >> For some reason I don't see that reply from Jonathan in my Inbox. I'll >> try to google it. >> >> What should be my next step in that case? I can't use pig then? >> >> On Thu, May 26, 2011 at 10:00 AM, Harsh J <ha...@cloudera.com> wrote: >> > I think Jonathan Coveney's reply on user@pig answered your question. >> > Its basically an issue of hadoop version differences between the one >> > Pig 0.8.1 release got bundled with vs. Hadoop 0.20.203 release which >> > is newer. >> > >> > On Thu, May 26, 2011 at 10:26 PM, Mohit Anchlia <mohitanch...@gmail.com> >> wrote: >> >> I sent this to pig apache user mailing list but have got no response. >> >> Not sure if that list is still active. >> >> >> >> thought I will post here if someone is able to help me. >> >> >> >> I am in process of installing and learning pig. I have a hadoop >> >> cluster and when I try to run pig in mapreduce mode it errors out: >> >> >> >> Hadoop version is hadoop-0.20.203.0 and pig version is pig-0.8.1 >> >> >> >> Error before Pig is launched >> >> ---------------------------- >> >> ERROR 2999: Unexpected internal error. Failed to create DataStorage >> >> >> >> java.lang.RuntimeException: Failed to create DataStorage >> >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:75) >> >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.<init>(HDataStorage.java:58) >> >> at >> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:214) >> >> at >> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:134) >> >> at org.apache.pig.impl.PigContext.connect(PigContext.java:183) >> >> at org.apache.pig.PigServer.<init>(PigServer.java:226) >> >> at org.apache.pig.PigServer.<init>(PigServer.java:215) >> >> at org.apache.pig.tools.grunt.Grunt.<init>(Grunt.java:55) >> >> at org.apache.pig.Main.run(Main.java:452) >> >> at org.apache.pig.Main.main(Main.java:107) >> >> Caused by: java.io.IOException: Call to dsdb1/172.18.60.96:54310 >> >> failed on local exception: java.io.EOFException >> >> at org.apache.hadoop.ipc.Client.wrapException(Client.java:775) >> >> at org.apache.hadoop.ipc.Client.call(Client.java:743) >> >> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >> >> at $Proxy0.getProtocolVersion(Unknown Source) >> >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) >> >> at >> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106) >> >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207) >> >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170) >> >> at >> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82) >> >> at >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378) >> >> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >> >> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390) >> >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196) >> >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95) >> >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:72) >> >> ... 9 more >> >> Caused by: java.io.EOFException >> >> at java.io.DataInputStream.readInt(DataInputStream.java:375) >> >> at >> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501) >> >> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446) >> >> >> > >> > >> > >> > -- >> > Harsh J >> > >> >