For some reason I don't see that reply from Jonathan in my Inbox. I'll
try to google it.

What should be my next step in that case? I can't use pig then?

On Thu, May 26, 2011 at 10:00 AM, Harsh J <ha...@cloudera.com> wrote:
> I think Jonathan Coveney's reply on user@pig answered your question.
> Its basically an issue of hadoop version differences between the one
> Pig 0.8.1 release got bundled with vs. Hadoop 0.20.203 release which
> is newer.
>
> On Thu, May 26, 2011 at 10:26 PM, Mohit Anchlia <mohitanch...@gmail.com> 
> wrote:
>> I sent this to pig apache user mailing list but have got no response.
>> Not sure if that list is still active.
>>
>> thought I will post here if someone is able to help me.
>>
>> I am in process of installing and learning pig. I have a hadoop
>> cluster and when I try to run pig in mapreduce mode it errors out:
>>
>> Hadoop version is hadoop-0.20.203.0 and pig version is pig-0.8.1
>>
>> Error before Pig is launched
>> ----------------------------
>> ERROR 2999: Unexpected internal error. Failed to create DataStorage
>>
>> java.lang.RuntimeException: Failed to create DataStorage
>>       at 
>> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:75)
>>       at 
>> org.apache.pig.backend.hadoop.datastorage.HDataStorage.<init>(HDataStorage.java:58)
>>       at 
>> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:214)
>>       at 
>> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:134)
>>       at org.apache.pig.impl.PigContext.connect(PigContext.java:183)
>>       at org.apache.pig.PigServer.<init>(PigServer.java:226)
>>       at org.apache.pig.PigServer.<init>(PigServer.java:215)
>>       at org.apache.pig.tools.grunt.Grunt.<init>(Grunt.java:55)
>>       at org.apache.pig.Main.run(Main.java:452)
>>       at org.apache.pig.Main.main(Main.java:107)
>> Caused by: java.io.IOException: Call to dsdb1/172.18.60.96:54310
>> failed on local exception: java.io.EOFException
>>       at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
>>       at org.apache.hadoop.ipc.Client.call(Client.java:743)
>>       at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
>>       at $Proxy0.getProtocolVersion(Unknown Source)
>>       at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
>>       at 
>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106)
>>       at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207)
>>       at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170)
>>       at 
>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82)
>>       at 
>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)
>>       at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
>>       at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)
>>       at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)
>>       at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95)
>>       at 
>> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:72)
>>       ... 9 more
>> Caused by: java.io.EOFException
>>       at java.io.DataInputStream.readInt(DataInputStream.java:375)
>>       at 
>> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501)
>>       at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446)
>>
>
>
>
> --
> Harsh J
>

Reply via email to