Adding -Dhadoop.version="0.18.2" to the ant build is not sufficient on its own.
ALSO: Run ant clean Clean out the target dir then Build with the right hadoop version flag passed in. If you switch the hadoop version to build, you must run an ant clean for the change to take effect. On 3/4/09 3:17 AM, "Marcus Herou" <marcus.he...@tailsweep.com> wrote: Hi Johan and thanks for the fast reply. It seemed like Hive did not find the method yes. That did not do it, same exception... I wiped the dist dir and dropped/created/loaded the data then issued the query again. Should I as well wipe some other local/hdfs dir ? Upgrading to 0.19.1 is not an option for some weeks. On Wed, Mar 4, 2009 at 11:59 AM, Johan Oskarsson <jo...@oskarsson.nu> wrote: Hi Marcus, It looks like you've hit on a Hadoop 0.18 vs 0.19 issue, try to compile Hive using: ant -Dhadoop.version="0.18.2" package That runs some preprocessing steps to remove 0.19 specific code from Hive. /Johan Marcus Herou wrote: > Hi. > > Started to lab with Hive today since it seems to suit us quite well and > since we are processing our weblogstats with Hadoop as of today and ends > up doing SQL in Hadoop form it seems fair to try out a system which does > that in one step :) > > I've created and loaded data into Hive with the following statements; > hive> drop table DailyUniqueSiteVisitorSample; > OK > Time taken: 4.064 seconds > hive> CREATE TABLE DailyUniqueSiteVisitorSample (sampleDate date,uid > bigint,site int,concreteStatistics int,network smallint,category > smallint,country smallint,countryCode String,sessions > smallint,pageImpressions smallint) COMMENT 'This is our weblog stats > table' PARTITIONED BY(dt STRING) ROW FORMAT DELIMITED FIELDS TERMINATED > BY ',' LINES TERMINATED BY '\n' STORED AS TEXTFILE; > OK > Time taken: 0.248 seconds > hive> LOAD DATA LOCAL INPATH > '/tmp/data-DenormalizedSiteVisitor.VisitsPi.2009-03-02.csv' INTO TABLE > DailyUniqueSiteVisitorSample PARTITION(dt='2009-03-02'); > Copying data from file:/tmp/data-2009-03-02.csv > Loading data to table dailyuniquesitevisitorsample partition {dt=2009-03-02} > OK > Time taken: 2.258 seconds > > A little confused about the text-file part but since the csv I need to > insert is a text-file so... (the tutorial only uses SequenceFile(s)), > seems to work though. > > Anyway this goes well but when I issue a simple query like the below it > throws an exception: > hive> select d.* from dailyuniquesitevisitorsample d where d.site=1; > Total MapReduce jobs = 1 > Number of reduce tasks is set to 0 since there's no reduce operator > java.lang.AbstractMethodError: > org.apache.hadoop.hive.ql.io.HiveInputFormat.validateInput(Lorg/apache/hadoop/mapred/JobConf;)V > at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:735) > at > org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:391) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:239) > at > org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174) > at > org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:306) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at org.apache.hadoop.util.RunJar.main(RunJar.java:155) > at org.apache.hadoop.mapred.JobShell.run(JobShell.java:54) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > at org.apache.hadoop.mapred.JobShell.main(JobShell.java:68) > > I run Hadoop-018.2 > > Not sure that I am doing this correctly. Please guide me if I am stupid. > > Kindly > > //Marcus > > > -- > Marcus Herou CTO and co-founder Tailsweep AB > +46702561312 > marcus.he...@tailsweep.com <mailto:marcus.he...@tailsweep.com> > http://www.tailsweep.com/ > http://blogg.tailsweep.com/