Hi Josh. Ok, got it. Interesting.
Downloaded ant, recompiled and now it works. Thank you. On Fri, Jan 10, 2014 at 10:16 PM, Josh Elser <josh.el...@gmail.com> wrote: > Mariano, > > Pig 0.12.0 does work with Hadoop-2.2.0, but you need to recompile Pig > first. > > In your $PIG_HOME, run the following command to rebuild Pig: > > `ant clean jar-withouthadoop -Dhadoopversion=23` > > Then, try re-running your script. > > > On 1/9/14, 5:13 PM, Mariano Kamp wrote: > >> Hi, >> >> I am trying to run the first pig sample from the Hadoop - The definitive >> guide - book. >> >> Unfortunately that doesn't work for me. >> >> I downloaded 0.12.0 and got the impression it should work with Hadoop 2.2. >> >> http://pig.apache.org/releases.html#14+October%2C+ >>> 2013%3A+release+0.12.0+available >>> 14 October, 2013: release 0.12.0 available >>> This release include several new features such as ASSERT operator, >>> Streaming UDF, new AvroStorage, IN/CASE >operator, BigInteger/BigDecimal >>> data type, support for Windows. >>> Note >>> This release works with Hadoop 0.20.X, 1.X, 0.23.X and 2.X >>> >> >> I use Hadoop 2.x. >> >>> snow:bin mkamp$ which hadoop >>> /Users/mkamp/hadoop-2.2.0/bin//hadoop >>> >> >> snow:bin mkamp$ echo $HADOOP_HOME >>> /Users/mkamp/hadoop-2.2.0 >>> >> >> But no matter if HADOOP_HOME is set or not I get a couple of errors and >> it doesn't work if I run the script: >> >> records = LOAD 'micro-tab/sample.txt' >>> AS (year:chararray, temperature:int, quality:int); >>> DUMP records; >>> >> >> All hell breaks lose and there is a lot of output, but most seems >> meaningless, warnings about settings that are deprecated in Hadoop, but >> still delivered by default this way. >> >> Hard to say what is relevant. Here are some excerpts, full output >> attached as file. >> >> From the logfile: >> >> Unexpected System Error Occured: java.lang.IncompatibleClassChangeError: >>> Found interface >org.apache.hadoop.mapreduce.JobContext, but class was >>> expected >>> at >org.apache.pig.backend.hadoop.executionengine.mapReduceLayer. >>> PigOutputFormat.setupUdfEnvAndStores(PigO>utputFormat.java:225) >>> >> >> >> ERROR 1066: Unable to open iterator for alias records >>> From the console: >>> >> >> 2014-01-09 22:24:45,976 [main] WARN >> org.apache.pig.backend.hadoop20.PigJobControl >>> - falling back to default >JobControl (not using hadoop 0.20 ?) >>> java.lang.NoSuchFieldException: runnerState >>> at java.lang.Class.getDeclaredField(Class.java:1918) >>> >> >> But as a little googling indicated, this is business as usual? >> >> 2014-01-09 22:24:49,228 [JobControl] ERROR >> org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl >>> - Error >while trying to run jobs. >>> java.lang.IncompatibleClassChangeError: Found interface >>> org.apache.hadoop.mapreduce.JobContext, but class >was expected >>> >> >> Input(s): >>> Failed to read data from "hdfs://localhost/user/mkamp/ >>> micro-tab/sample.txt" >>> >> >> That last one looks interesting. Maybe I am using it wrong and the >> reported errors are not related? I wanted to read from the local file >> system. >> >> So I also changed the script to read from hdfs, but that didn't change >> the error. >> >> Any ideas where to go from here? >> >> Is it possible to run the latest Hadoop binary download and the latest >> Pig binary download together? >> >>