Margusja, >From trunk, can you build mahout using the following command and try again: mvn clean package -DskipTests=true -Dhadoop2.version=2.2.0
Best Gokhan On Tue, Mar 4, 2014 at 4:25 PM, Margusja <mar...@roo.ee> wrote: > Hi thanks for reply. > > Here is my output: > > [hduser@vm38 ~]$ /usr/lib/hadoop/bin/hadoop version Hadoop > 2.2.0.2.0.6.0-101 > Subversion g...@github.com:hortonworks/hadoop.git -r > b07b2906c36defd389c8b5bd22bebc1bead8115b > Compiled by jenkins on 2014-01-09T05:18Z > Compiled with protoc 2.5.0 > From source with checksum 704f1e463ebc4fb89353011407e965 > This command was run using /usr/lib/hadoop/hadoop-common- > 2.2.0.2.0.6.0-101.jar > > [hduser@vm38 ~]$ /usr/lib/hadoop/bin/hadoop jar > mahout/examples/target/mahout-examples-1.0-SNAPSHOT-job.jar > org.apache.mahout.classifier.df.mapreduce.BuildForest -d > input/data666.noheader.data -ds input/data666.noheader.data.info -sl 5 -p > -t 100 -o nsl-forest > > ... > 14/03/04 16:22:51 INFO mapreduce.Job: map 0% reduce 0% > 14/03/04 16:23:12 INFO mapreduce.Job: map 100% reduce 0% > 14/03/04 16:23:43 INFO mapreduce.Job: Job job_1393936067845_0013 completed > successfully > 14/03/04 16:23:44 INFO mapreduce.Job: Counters: 27 > > File System Counters > FILE: Number of bytes read=2994 > FILE: Number of bytes written=80677 > FILE: Number of read operations=0 > FILE: Number of large read operations=0 > FILE: Number of write operations=0 > HDFS: Number of bytes read=880103 > HDFS: Number of bytes written=2436546 > > HDFS: Number of read operations=5 > HDFS: Number of large read operations=0 > HDFS: Number of write operations=2 > Job Counters > Launched map tasks=1 > Data-local map tasks=1 > Total time spent by all maps in occupied slots (ms)=45253 > > Total time spent by all reduces in occupied slots (ms)=0 > Map-Reduce Framework > Map input records=9994 > Map output records=100 > Input split bytes=123 > Spilled Records=0 > Failed Shuffles=0 > Merged Map outputs=0 > GC time elapsed (ms)=456 > CPU time spent (ms)=36010 > Physical memory (bytes) snapshot=180752384 > Virtual memory (bytes) snapshot=994275328 > Total committed heap usage (bytes)=101187584 > > File Input Format Counters > Bytes Read=879980 > File Output Format Counters > Bytes Written=2436546 > > Exception in thread "main" java.lang.IncompatibleClassChangeError: Found > interface org.apache.hadoop.mapreduce.JobContext, but class was expected > at org.apache.mahout.classifier.df.mapreduce.partial. > PartialBuilder.processOutput(PartialBuilder.java:113) > at org.apache.mahout.classifier.df.mapreduce.partial. > PartialBuilder.parseOutput(PartialBuilder.java:89) > at org.apache.mahout.classifier.df.mapreduce.Builder.build( > Builder.java:294) > at org.apache.mahout.classifier.df.mapreduce.BuildForest. > buildForest(BuildForest.java:228) > at org.apache.mahout.classifier.df.mapreduce.BuildForest.run( > BuildForest.java:188) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) > at org.apache.mahout.classifier.df.mapreduce.BuildForest.main( > BuildForest.java:252) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:606) > at org.apache.hadoop.util.RunJar.main(RunJar.java:212) > > > > Tervitades, Margus (Margusja) Roo > +372 51 48 780 > http://margus.roo.ee > http://ee.linkedin.com/in/margusroo > skype: margusja > ldapsearch -x -h ldap.sk.ee -b c=EE "(serialNumber=37303140314)" > -----BEGIN PUBLIC KEY----- > MIGfMA0GCSqGSIb3DQEBAQUAA4GNADCBiQKBgQCvbeg7LwEC2SCpAEewwpC3ajxE > 5ZsRMCB77L8bae9G7TslgLkoIzo9yOjPdx2NN6DllKbV65UjTay43uUDyql9g3tl > RhiJIcoAExkSTykWqAIPR88LfilLy1JlQ+0RD8OXiWOVVQfhOHpQ0R/jcAkM2lZa > BjM8j36yJvoBVsfOHQIDAQAB > -----END PUBLIC KEY----- > > On 04/03/14 16:11, Sergey Svinarchuk wrote: > >> Sory, I didn't see that you try use mahout-1.0-snapshot. >> You used /usr/lib/hadoop-yarn/bin/yarn but need use >> /usr/lib/hadoop/bin/hadoop and then your example will be success. >> >> >> On Tue, Mar 4, 2014 at 3:45 PM, Sergey Svinarchuk < >> ssvinarc...@hortonworks.com> wrote: >> >> Mahout 0.9 not supported hadoop 2 dependencies. >>> You can use mahout-1.0-SNAPSHOT or add to your mahout patch from >>> https://issues.apache.org/jira/browse/MAHOUT-1329 for added hadoop 2 >>> support. >>> >>> >>> On Tue, Mar 4, 2014 at 3:38 PM, Margusja <mar...@roo.ee> wrote: >>> >>> Hi >>>> >>>> following command: >>>> /usr/lib/hadoop-yarn/bin/yarn jar mahout-distribution-0.9/ >>>> mahout-examples-0.9.jar >>>> org.apache.mahout.classifier.df.mapreduce.BuildForest -d >>>> input/data666.noheader.data -ds input/data666.noheader.data.info -sl 5 >>>> -p -t 100 -o nsl-forest >>>> >>>> When I used hadoop 1.x then it worked. >>>> Now I use hadoop-2.2.0 it gives me: >>>> 14/03/04 15:25:58 INFO mapreduce.BuildForest: Partial Mapred >>>> implementation >>>> 14/03/04 15:25:58 INFO mapreduce.BuildForest: Building the forest... >>>> 14/03/04 15:26:01 INFO client.RMProxy: Connecting to ResourceManager at >>>> / >>>> 0.0.0.0:8032 >>>> 14/03/04 15:26:05 INFO input.FileInputFormat: Total input paths to >>>> process : 1 >>>> 14/03/04 15:26:05 INFO mapreduce.JobSubmitter: number of splits:1 >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: user.name is >>>> deprecated. Instead, use mapreduce.job.user.name >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: mapred.jar is >>>> deprecated. Instead, use mapreduce.job.jar >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: >>>> mapred.cache.files.filesizes is deprecated. Instead, use >>>> mapreduce.job.cache.files.filesizes >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: mapred.cache.files is >>>> deprecated. Instead, use mapreduce.job.cache.files >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: mapred.reduce.tasks is >>>> deprecated. Instead, use mapreduce.job.reduces >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: >>>> mapred.output.value.class is deprecated. Instead, use >>>> mapreduce.job.output.value.class >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: mapreduce.map.class is >>>> deprecated. Instead, use mapreduce.job.map.class >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: mapred.job.name is >>>> deprecated. Instead, use mapreduce.job.name >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: >>>> mapreduce.inputformat.class is deprecated. Instead, use >>>> mapreduce.job.inputformat.class >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: mapred.input.dir is >>>> deprecated. Instead, use mapreduce.input.fileinputformat.inputdir >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: mapred.output.dir is >>>> deprecated. Instead, use mapreduce.output.fileoutputformat.outputdir >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: >>>> mapreduce.outputformat.class is deprecated. Instead, use >>>> mapreduce.job.outputformat.class >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: mapred.map.tasks is >>>> deprecated. Instead, use mapreduce.job.maps >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: >>>> mapred.cache.files.timestamps is deprecated. Instead, use >>>> mapreduce.job.cache.files.timestamps >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: >>>> mapred.output.key.class >>>> is deprecated. Instead, use mapreduce.job.output.key.class >>>> 14/03/04 15:26:05 INFO Configuration.deprecation: mapred.working.dir is >>>> deprecated. Instead, use mapreduce.job.working.dir >>>> 14/03/04 15:26:06 INFO mapreduce.JobSubmitter: Submitting tokens for >>>> job: >>>> job_1393936067845_0011 >>>> 14/03/04 15:26:07 INFO impl.YarnClientImpl: Submitted application >>>> application_1393936067845_0011 to ResourceManager at /0.0.0.0:8032 >>>> 14/03/04 15:26:07 INFO mapreduce.Job: The url to track the job: >>>> http://vm38.dbweb.ee:8088/proxy/application_1393936067845_0011/ >>>> 14/03/04 15:26:07 INFO mapreduce.Job: Running job: >>>> job_1393936067845_0011 >>>> 14/03/04 15:26:36 INFO mapreduce.Job: Job job_1393936067845_0011 running >>>> in uber mode : false >>>> 14/03/04 15:26:36 INFO mapreduce.Job: map 0% reduce 0% >>>> 14/03/04 15:27:00 INFO mapreduce.Job: map 100% reduce 0% >>>> 14/03/04 15:27:26 INFO mapreduce.Job: Job job_1393936067845_0011 >>>> completed successfully >>>> 14/03/04 15:27:26 INFO mapreduce.Job: Counters: 27 >>>> File System Counters >>>> FILE: Number of bytes read=2994 >>>> FILE: Number of bytes written=80677 >>>> FILE: Number of read operations=0 >>>> FILE: Number of large read operations=0 >>>> FILE: Number of write operations=0 >>>> HDFS: Number of bytes read=880103 >>>> HDFS: Number of bytes written=2483042 >>>> HDFS: Number of read operations=5 >>>> HDFS: Number of large read operations=0 >>>> HDFS: Number of write operations=2 >>>> Job Counters >>>> Launched map tasks=1 >>>> Data-local map tasks=1 >>>> Total time spent by all maps in occupied slots >>>> (ms)=46056 >>>> Total time spent by all reduces in occupied slots >>>> (ms)=0 >>>> Map-Reduce Framework >>>> Map input records=9994 >>>> Map output records=100 >>>> Input split bytes=123 >>>> Spilled Records=0 >>>> Failed Shuffles=0 >>>> Merged Map outputs=0 >>>> GC time elapsed (ms)=425 >>>> CPU time spent (ms)=32890 >>>> Physical memory (bytes) snapshot=189755392 >>>> Virtual memory (bytes) snapshot=992145408 >>>> Total committed heap usage (bytes)=111673344 >>>> File Input Format Counters >>>> Bytes Read=879980 >>>> File Output Format Counters >>>> Bytes Written=2483042 >>>> Exception in thread "main" java.lang.IncompatibleClassChangeError: >>>> Found >>>> interface org.apache.hadoop.mapreduce.JobContext, but class was >>>> expected >>>> at org.apache.mahout.classifier.df.mapreduce.partial. >>>> PartialBuilder.processOutput(PartialBuilder.java:113) >>>> at org.apache.mahout.classifier.df.mapreduce.partial. >>>> PartialBuilder.parseOutput(PartialBuilder.java:89) >>>> at org.apache.mahout.classifier.df.mapreduce.Builder.build( >>>> Builder.java:294) >>>> at org.apache.mahout.classifier.df.mapreduce.BuildForest. >>>> buildForest(BuildForest.java:228) >>>> at org.apache.mahout.classifier.df.mapreduce.BuildForest.run( >>>> BuildForest.java:188) >>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>> at org.apache.mahout.classifier.df.mapreduce.BuildForest.main( >>>> BuildForest.java:252) >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> at sun.reflect.NativeMethodAccessorImpl.invoke( >>>> NativeMethodAccessorImpl.java:57) >>>> at sun.reflect.DelegatingMethodAccessorImpl.invoke( >>>> DelegatingMethodAccessorImpl.java:43) >>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:212) >>>> >>>> I even downloaded source from https://github.com/apache/mahout.git and >>>> build it like: >>>> mvn -DskipTests -Dhadoop2.version=2.2.0 clean install >>>> then used command line: >>>> /usr/lib/hadoop-yarn/bin/yarn jar mahout/examples/target/mahout- >>>> examples-1.0-SNAPSHOT-job.jar >>>> org.apache.mahout.classifier.df.mapreduce.BuildForest -d >>>> input/data666.noheader.data -ds input/data666.noheader.data.info -sl 5 >>>> -p -t 100 -o nsl-forest >>>> >>>> and got the same error like above. >>>> >>>> Is there something wrong in my side or hadoop-2.2.0 and mahout can not >>>> play each other anymore? >>>> >>>> The typical example: >>>> /usr/lib/hadoop-yarn/bin/yarn jar /usr/lib/hadoop-mapreduce/ >>>> hadoop-mapreduce-examples-2.2.0.2.0.6.0-101.jar pi 2 5 >>>> works. >>>> >>>> -- >>>> Tervitades, Margus (Margusja) Roo >>>> +372 51 48 780 >>>> http://margus.roo.ee >>>> http://ee.linkedin.com/in/margusroo >>>> skype: margusja >>>> ldapsearch -x -h ldap.sk.ee -b c=EE "(serialNumber=37303140314)" >>>> -----BEGIN PUBLIC KEY----- >>>> MIGfMA0GCSqGSIb3DQEBAQUAA4GNADCBiQKBgQCvbeg7LwEC2SCpAEewwpC3ajxE >>>> 5ZsRMCB77L8bae9G7TslgLkoIzo9yOjPdx2NN6DllKbV65UjTay43uUDyql9g3tl >>>> RhiJIcoAExkSTykWqAIPR88LfilLy1JlQ+0RD8OXiWOVVQfhOHpQ0R/jcAkM2lZa >>>> BjM8j36yJvoBVsfOHQIDAQAB >>>> -----END PUBLIC KEY----- >>>> >>>> >>>> >