Trouble in running MapReduce application
Hi everyone, I know this is the common mistake to not specify the class adress while trying to run a jar, however, although I specified, I am still getting the ClassNotFound exception. What may be the reason for it? I have been struggling for this problem more than a 2 days. I just wrote different MapReduce application for some anlaysis. I got this problem. To check, is there something wrong with my system, i tried to run WordCount example. When I just run hadoop-examples wordcount, it is working fine. But when I add just "package org.myorg;" command at the beginning, it doesnot work. Here is what I have done so far * 1. I just copied wordcount code from the apaches own examples source code and I just changed package decleration as "package org.myorg;" ** 2. Then I tried to run that command: * "hadoop jar wordcount_19_02.jar org.myorg.WordCount /home/hadoop/project/hadoop-data/NetFlow 19_02_wordcount.output" * 3. I got following error: ** [hadoop@ADUAE042-LAP-V project]$ hadoop jar wordcount_19_02.jar org.myorg.WordCount /home/hadoop/project/hadoop-data/NetFlow 19_02_wordcount.output Warning: $HADOOP_HOME is deprecated. Exception in thread "main" java.lang.ClassNotFoundException: org.myorg.WordCount at java.net.URLClassLoader$1.run(URLClassLoader.java:217) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:205) at java.lang.ClassLoader.loadClass(ClassLoader.java:321) at java.lang.ClassLoader.loadClass(ClassLoader.java:266) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:266) at org.apache.hadoop.util.RunJar.main(RunJar.java:149) ** 4. This is the content of my .jar file: [hadoop@ADUAE042-LAP-V project]$ jar tf wordcount_19_02.jar META-INF/ META-INF/MANIFEST.MF wordcount_classes/ wordcount_classes/org/ wordcount_classes/org/myorg/ wordcount_classes/org/myorg/WordCount.class wordcount_classes/org/myorg/WordCount$TokenizerMapper.class wordcount_classes/org/myorg/WordCount$IntSumReducer.class ** 5. This is the 'ls' output of my working directory: ** [hadoop@ADUAE042-LAP-V project]$ ls flowclasses_18_02 flowclasses_18_02.jar hadoop-1.0.4 hadoop-1.0.4.tar.gz hadoop-data MapReduce.java sample wordcount_19_02.jar wordcount_classes WordCountClasses WordCount.java * So as you see, package decleration is fine but I am really helpless, I googled but they are all saying samething you should specify the package hierarchy of your main class. I did know it already I am specifying but doesn't work. I would be much obliged to anyone helped me Regards,
Re: Trouble in running MapReduce application
Your point (4) explains the problem. The jar packed structure should look like the below, and not how it is presently (one extra top level dir is present): META-INF/ META-INF/MANIFEST.MF org/ org/myorg/ org/myorg/WordCount.class org/myorg/WordCount$TokenizerMapper.class org/myorg/WordCount$IntSumReducer.class On Tue, Feb 19, 2013 at 9:29 PM, Fatih Haltas wrote: > Hi everyone, > > I know this is the common mistake to not specify the class adress while > trying to run a jar, however, > although I specified, I am still getting the ClassNotFound exception. > > What may be the reason for it? I have been struggling for this problem more > than a 2 days. > I just wrote different MapReduce application for some anlaysis. I got this > problem. > > To check, is there something wrong with my system, i tried to run WordCount > example. > When I just run hadoop-examples wordcount, it is working fine. > > But when I add just "package org.myorg;" command at the beginning, it > doesnot work. > > Here is what I have done so far > * > 1. I just copied wordcount code from the apaches own examples source code > and I just changed package decleration as "package org.myorg;" > ** > 2. Then I tried to run that command: > * > "hadoop jar wordcount_19_02.jar org.myorg.WordCount > /home/hadoop/project/hadoop-data/NetFlow 19_02_wordcount.output" > * > 3. I got following error: > ** > [hadoop@ADUAE042-LAP-V project]$ hadoop jar wordcount_19_02.jar > org.myorg.WordCount /home/hadoop/project/hadoop-data/NetFlow > 19_02_wordcount.output > Warning: $HADOOP_HOME is deprecated. > > Exception in thread "main" java.lang.ClassNotFoundException: > org.myorg.WordCount > at java.net.URLClassLoader$1.run(URLClassLoader.java:217) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:205) > at java.lang.ClassLoader.loadClass(ClassLoader.java:321) > at java.lang.ClassLoader.loadClass(ClassLoader.java:266) > at java.lang.Class.forName0(Native Method) > at java.lang.Class.forName(Class.java:266) > at org.apache.hadoop.util.RunJar.main(RunJar.java:149) > > ** > 4. This is the content of my .jar file: > > [hadoop@ADUAE042-LAP-V project]$ jar tf wordcount_19_02.jar > META-INF/ > META-INF/MANIFEST.MF > wordcount_classes/ > wordcount_classes/org/ > wordcount_classes/org/myorg/ > wordcount_classes/org/myorg/WordCount.class > wordcount_classes/org/myorg/WordCount$TokenizerMapper.class > wordcount_classes/org/myorg/WordCount$IntSumReducer.class > ** > 5. This is the 'ls' output of my working directory: > ** > [hadoop@ADUAE042-LAP-V project]$ ls > flowclasses_18_02 flowclasses_18_02.jar hadoop-1.0.4 hadoop-1.0.4.tar.gz > hadoop-data MapReduce.java sample wordcount_19_02.jar wordcount_classes > WordCountClasses WordCount.java > * > So as you see, package decleration is fine but I am really helpless, I > googled but they are all saying samething you should specify the package > hierarchy of your main class. I did know it already I am specifying but > doesn't work. > > I would be much obliged to anyone helped me > > Regards, -- Harsh J
Re: Trouble in running MapReduce application
Thank you very much Harsh, Now, as I promised earlier I am much obliged to you. But, now I solved that problem by just changing the directories then again creating a jar file of org. but I am getting this error: 1.) What I got -- [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ hadoop jar flow19028pm.jar org.myorg.MapReduce /home/hadoop/project/hadoop-data/NetFlow 19_02.out Warning: $HADOOP_HOME is deprecated. Exception in thread "main" java.lang.UnsupportedClassVersionError: org/myorg/MapReduce : Unsupported major.minor version 51.0 at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:634) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) at java.net.URLClassLoader.access$000(URLClassLoader.java:73) at java.net.URLClassLoader$1.run(URLClassLoader.java:212) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:205) at java.lang.ClassLoader.loadClass(ClassLoader.java:321) at java.lang.ClassLoader.loadClass(ClassLoader.java:266) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:266) at org.apache.hadoop.util.RunJar.main(RunJar.java:149) 2.) How I create my jar - [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar cvf flow19028pm.jar org added manifest adding: org/(in = 0) (out= 0)(stored 0%) adding: org/myorg/(in = 0) (out= 0)(stored 0%) adding: org/myorg/MapReduce$FlowPortReducer.class(in = 1661) (out= 690)(deflated 58%) adding: org/myorg/MapReduce.class(in = 1587) (out= 903)(deflated 43%) adding: org/myorg/MapReduce$FlowPortMapper.class(in = 1874) (out= 823)(deflated 56%) 3.) Content of my jar file --- [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar tf flow19028pm.jar META-INF/ META-INF/MANIFEST.MF org/ org/myorg/ org/myorg/MapReduce$FlowPortReducer.class org/myorg/MapReduce.class org/myorg/MapReduce$FlowPortMapper.class - Thank you very much. On Tue, Feb 19, 2013 at 8:20 PM, Harsh J wrote: > Your point (4) explains the problem. The jar packed structure should > look like the below, and not how it is presently (one extra top level > dir is present): > > META-INF/ > META-INF/MANIFEST.MF > org/ > org/myorg/ > org/myorg/WordCount.class > org/myorg/WordCount$TokenizerMapper.class > org/myorg/WordCount$IntSumReducer.class > > On Tue, Feb 19, 2013 at 9:29 PM, Fatih Haltas > wrote: > > Hi everyone, > > > > I know this is the common mistake to not specify the class adress while > > trying to run a jar, however, > > although I specified, I am still getting the ClassNotFound exception. > > > > What may be the reason for it? I have been struggling for this problem > more > > than a 2 days. > > I just wrote different MapReduce application for some anlaysis. I got > this > > problem. > > > > To check, is there something wrong with my system, i tried to run > WordCount > > example. > > When I just run hadoop-examples wordcount, it is working fine. > > > > But when I add just "package org.myorg;" command at the beginning, it > > doesnot work. > > > > Here is what I have done so far > > * > > 1. I just copied wordcount code from the apaches own examples source code > > and I just changed package decleration as "package org.myorg;" > > > ** > > 2. Then I tried to run that command: > > > * > > "hadoop jar wordcount_19_02.jar org.myorg.WordCount > > /home/hadoop/project/hadoop-data/NetFlow 19_02_wordcount.output" > > * > > 3. I got following error: > > > ** > > [hadoop@ADUAE042-LAP-V project]$ hadoop jar wordcount_19_02.jar > > org.myorg.WordCount /home/hadoop/project/hadoop-data/NetFlow > > 19_02_wordcount.output > > Warning: $HADOOP_HOME is deprecated. > > > > Exception in thread "main" java.lang.ClassNotFoundException: > > org.myorg.WordCount > > at java.net.URLClassLoader$1.run(URLClassLoader.java:217) > > at java.security.AccessController.doPrivileged(Native Method) > > at java.net.URLClassLoader.findClass(URLClassLoader.java:205) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:321) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:266) > > a
Re: Trouble in running MapReduce application
Hi, The new error usually happens if you compile using Java 7 and try to run via Java 6 (for example). That is, an incompatibility in the runtimes for the binary artifact produced. On Tue, Feb 19, 2013 at 10:09 PM, Fatih Haltas wrote: > Thank you very much Harsh, > > Now, as I promised earlier I am much obliged to you. > > But, now I solved that problem by just changing the directories then again > creating a jar file of org. but I am getting this error: > > 1.) What I got > -- > [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ hadoop jar flow19028pm.jar > org.myorg.MapReduce /home/hadoop/project/hadoop-data/NetFlow 19_02.out > Warning: $HADOOP_HOME is deprecated. > > Exception in thread "main" java.lang.UnsupportedClassVersionError: > org/myorg/MapReduce : Unsupported major.minor version 51.0 > at java.lang.ClassLoader.defineClass1(Native Method) > at java.lang.ClassLoader.defineClass(ClassLoader.java:634) > at > java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) > at java.net.URLClassLoader.access$000(URLClassLoader.java:73) > at java.net.URLClassLoader$1.run(URLClassLoader.java:212) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:205) > at java.lang.ClassLoader.loadClass(ClassLoader.java:321) > at java.lang.ClassLoader.loadClass(ClassLoader.java:266) > at java.lang.Class.forName0(Native Method) > at java.lang.Class.forName(Class.java:266) > at org.apache.hadoop.util.RunJar.main(RunJar.java:149) > > 2.) How I create my jar > - > [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar cvf flow19028pm.jar org > added manifest > adding: org/(in = 0) (out= 0)(stored 0%) > adding: org/myorg/(in = 0) (out= 0)(stored 0%) > adding: org/myorg/MapReduce$FlowPortReducer.class(in = 1661) (out= > 690)(deflated 58%) > adding: org/myorg/MapReduce.class(in = 1587) (out= 903)(deflated 43%) > adding: org/myorg/MapReduce$FlowPortMapper.class(in = 1874) (out= > 823)(deflated 56%) > > 3.) Content of my jar file > --- > [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar tf flow19028pm.jar > META-INF/ > META-INF/MANIFEST.MF > org/ > org/myorg/ > org/myorg/MapReduce$FlowPortReducer.class > org/myorg/MapReduce.class > org/myorg/MapReduce$FlowPortMapper.class > - > > > Thank you very much. > > > On Tue, Feb 19, 2013 at 8:20 PM, Harsh J wrote: >> >> Your point (4) explains the problem. The jar packed structure should >> look like the below, and not how it is presently (one extra top level >> dir is present): >> >> META-INF/ >> META-INF/MANIFEST.MF >> org/ >> org/myorg/ >> org/myorg/WordCount.class >> org/myorg/WordCount$TokenizerMapper.class >> org/myorg/WordCount$IntSumReducer.class >> >> On Tue, Feb 19, 2013 at 9:29 PM, Fatih Haltas >> wrote: >> > Hi everyone, >> > >> > I know this is the common mistake to not specify the class adress while >> > trying to run a jar, however, >> > although I specified, I am still getting the ClassNotFound exception. >> > >> > What may be the reason for it? I have been struggling for this problem >> > more >> > than a 2 days. >> > I just wrote different MapReduce application for some anlaysis. I got >> > this >> > problem. >> > >> > To check, is there something wrong with my system, i tried to run >> > WordCount >> > example. >> > When I just run hadoop-examples wordcount, it is working fine. >> > >> > But when I add just "package org.myorg;" command at the beginning, it >> > doesnot work. >> > >> > Here is what I have done so far >> > >> > * >> > 1. I just copied wordcount code from the apaches own examples source >> > code >> > and I just changed package decleration as "package org.myorg;" >> > >> > ** >> > 2. Then I tried to run that command: >> > >> > * >> > "hadoop jar wordcount_19_02.jar org.myorg.WordCount >> > /home/hadoop/project/hadoop-data/NetFlow 19_02_wordcount.output" >> > >> > * >> > 3. I got following error: >> > >> > ** >> > [hadoop@ADUAE042-LAP-V project]$ hadoop jar wordcount_19_02.jar >> > org.myorg.WordCount /home/hadoop/project/hadoop-data/NetFlow >> > 19_02_wordcount.output >> > Warning: $HADOOP_HOME is deprecated. >> > >> > Exception in thread "mai
Re: Trouble in running MapReduce application
Oops. I just noticed Hemanth has been answering on a dupe thread as well. Lets drop this thread and carry on there :) On Tue, Feb 19, 2013 at 11:14 PM, Harsh J wrote: > Hi, > > The new error usually happens if you compile using Java 7 and try to > run via Java 6 (for example). That is, an incompatibility in the > runtimes for the binary artifact produced. > > On Tue, Feb 19, 2013 at 10:09 PM, Fatih Haltas wrote: >> Thank you very much Harsh, >> >> Now, as I promised earlier I am much obliged to you. >> >> But, now I solved that problem by just changing the directories then again >> creating a jar file of org. but I am getting this error: >> >> 1.) What I got >> -- >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ hadoop jar flow19028pm.jar >> org.myorg.MapReduce /home/hadoop/project/hadoop-data/NetFlow 19_02.out >> Warning: $HADOOP_HOME is deprecated. >> >> Exception in thread "main" java.lang.UnsupportedClassVersionError: >> org/myorg/MapReduce : Unsupported major.minor version 51.0 >> at java.lang.ClassLoader.defineClass1(Native Method) >> at java.lang.ClassLoader.defineClass(ClassLoader.java:634) >> at >> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >> at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) >> at java.net.URLClassLoader.access$000(URLClassLoader.java:73) >> at java.net.URLClassLoader$1.run(URLClassLoader.java:212) >> at java.security.AccessController.doPrivileged(Native Method) >> at java.net.URLClassLoader.findClass(URLClassLoader.java:205) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:321) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:266) >> at java.lang.Class.forName0(Native Method) >> at java.lang.Class.forName(Class.java:266) >> at org.apache.hadoop.util.RunJar.main(RunJar.java:149) >> >> 2.) How I create my jar >> - >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar cvf flow19028pm.jar org >> added manifest >> adding: org/(in = 0) (out= 0)(stored 0%) >> adding: org/myorg/(in = 0) (out= 0)(stored 0%) >> adding: org/myorg/MapReduce$FlowPortReducer.class(in = 1661) (out= >> 690)(deflated 58%) >> adding: org/myorg/MapReduce.class(in = 1587) (out= 903)(deflated 43%) >> adding: org/myorg/MapReduce$FlowPortMapper.class(in = 1874) (out= >> 823)(deflated 56%) >> >> 3.) Content of my jar file >> --- >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar tf flow19028pm.jar >> META-INF/ >> META-INF/MANIFEST.MF >> org/ >> org/myorg/ >> org/myorg/MapReduce$FlowPortReducer.class >> org/myorg/MapReduce.class >> org/myorg/MapReduce$FlowPortMapper.class >> - >> >> >> Thank you very much. >> >> >> On Tue, Feb 19, 2013 at 8:20 PM, Harsh J wrote: >>> >>> Your point (4) explains the problem. The jar packed structure should >>> look like the below, and not how it is presently (one extra top level >>> dir is present): >>> >>> META-INF/ >>> META-INF/MANIFEST.MF >>> org/ >>> org/myorg/ >>> org/myorg/WordCount.class >>> org/myorg/WordCount$TokenizerMapper.class >>> org/myorg/WordCount$IntSumReducer.class >>> >>> On Tue, Feb 19, 2013 at 9:29 PM, Fatih Haltas >>> wrote: >>> > Hi everyone, >>> > >>> > I know this is the common mistake to not specify the class adress while >>> > trying to run a jar, however, >>> > although I specified, I am still getting the ClassNotFound exception. >>> > >>> > What may be the reason for it? I have been struggling for this problem >>> > more >>> > than a 2 days. >>> > I just wrote different MapReduce application for some anlaysis. I got >>> > this >>> > problem. >>> > >>> > To check, is there something wrong with my system, i tried to run >>> > WordCount >>> > example. >>> > When I just run hadoop-examples wordcount, it is working fine. >>> > >>> > But when I add just "package org.myorg;" command at the beginning, it >>> > doesnot work. >>> > >>> > Here is what I have done so far >>> > >>> > * >>> > 1. I just copied wordcount code from the apaches own examples source >>> > code >>> > and I just changed package decleration as "package org.myorg;" >>> > >>> > ** >>> > 2. Then I tried to run that command: >>> > >>> > * >>> > "hadoop jar wordcount_19_02.jar org.myorg.WordCount >>> > /home/hadoop/project/hadoop-data/NetFlow 19_02_wordcount.output" >>> > >>> > * >>> > 3. I got following error: >>> > >>> >
Re: Trouble in running MapReduce application
Thank you all very much 19 Şubat 2013 Salı tarihinde Harsh J adlı kullanıcı şöyle yazdı: > Oops. I just noticed Hemanth has been answering on a dupe thread as > well. Lets drop this thread and carry on there :) > > On Tue, Feb 19, 2013 at 11:14 PM, Harsh J wrote: > > Hi, > > > > The new error usually happens if you compile using Java 7 and try to > > run via Java 6 (for example). That is, an incompatibility in the > > runtimes for the binary artifact produced. > > > > On Tue, Feb 19, 2013 at 10:09 PM, Fatih Haltas > wrote: > >> Thank you very much Harsh, > >> > >> Now, as I promised earlier I am much obliged to you. > >> > >> But, now I solved that problem by just changing the directories then > again > >> creating a jar file of org. but I am getting this error: > >> > >> 1.) What I got > >> > -- > >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ hadoop jar flow19028pm.jar > >> org.myorg.MapReduce /home/hadoop/project/hadoop-data/NetFlow 19_02.out > >> Warning: $HADOOP_HOME is deprecated. > >> > >> Exception in thread "main" java.lang.UnsupportedClassVersionError: > >> org/myorg/MapReduce : Unsupported major.minor version 51.0 > >> at java.lang.ClassLoader.defineClass1(Native Method) > >> at java.lang.ClassLoader.defineClass(ClassLoader.java:634) > >> at > >> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > >> at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) > >> at java.net.URLClassLoader.access$000(URLClassLoader.java:73) > >> at java.net.URLClassLoader$1.run(URLClassLoader.java:212) > >> at java.security.AccessController.doPrivileged(Native Method) > >> at java.net.URLClassLoader.findClass(URLClassLoader.java:205) > >> at java.lang.ClassLoader.loadClass(ClassLoader.java:321) > >> at java.lang.ClassLoader.loadClass(ClassLoader.java:266) > >> at java.lang.Class.forName0(Native Method) > >> at java.lang.Class.forName(Class.java:266) > >> at org.apache.hadoop.util.RunJar.main(RunJar.java:149) > >> > >> 2.) How I create my jar > >> > - > >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar cvf flow19028pm.jar org > >> added manifest > >> adding: org/(in = 0) (out= 0)(stored 0%) > >> adding: org/myorg/(in = 0) (out= 0)(stored 0%) > >> adding: org/myorg/MapReduce$FlowPortReducer.class(in = 1661) (out= > >> 690)(deflated 58%) > >> adding: org/myorg/MapReduce.class(in = 1587) (out= 903)(deflated 43%) > >> adding: org/myorg/MapReduce$FlowPortMapper.class(in = 1874) (out= > >> 823)(deflated 56%) > >> > >> 3.) Content of my jar file > >> > --- > >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar tf flow19028pm.jar > >> META-INF/ > >> META-INF/MANIFEST.MF > >> org/ > >> org/myorg/ > >> org/myorg/MapReduce$FlowPortReducer.class > >> org/myorg/MapReduce.class > >> org/myorg/MapReduce$FlowPortMapper.class > >> > - > >> > >> > >> Thank you very much. > >> > >> > >> On Tue, Feb 19, 2013 at 8:20 PM, Harsh J wrote: > >>> > >>> Your point (4) explains the problem. The jar packed structure should > >>> look like the below, and not how it is presently (one extra top level > >>> dir is present): > >>> > >>> META-INF/ > >>> META-INF/MANIFEST.MF > >>> org/ > >>> org/myorg/ > >>> org/myorg/WordCount.class > >>> org/myorg/WordCount$TokenizerMapper.class > >>> org/myorg/WordCount$IntSumReducer.clas-- > Harsh J >
Re: Trouble in running MapReduce application
Hi again, Thanks for your help but now, I am struggling with the same problem on a machine. As the preivous problem, I just decrease the Java version by Java 6, but this time I could not solve the problem. those are outputs that may explain the situation: - 1. I could not run my own code, to check the system I just tried to run basic wordcount example without any modification, except package info. ** COMMAND EXECUTED: hadoop jar my.jar org.myorg.WordCount NetFlow NetFlow.out Warning: $HADOOP_HOME is deprecated. Exception in thread "main" java.lang.UnsupportedClassVersionError: org/myorg/WordCount : Unsupported major.minor version 51.0 at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:634) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) at java.net.URLClassLoader.access$000(URLClassLoader.java:73) at java.net.URLClassLoader$1.run(URLClassLoader.java:212) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:205) at java.lang.ClassLoader.loadClass(ClassLoader.java:321) at java.lang.ClassLoader.loadClass(ClassLoader.java:266) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:266) at org.apache.hadoop.util.RunJar.main(RunJar.java:149) ** 2. Java version: COMMAND EXECUTED: java -version java version "1.6.0_24" OpenJDK Runtime Environment (IcedTea6 1.11.6) (rhel-1.33.1.11.6.el5_9-x86_64) OpenJDK 64-Bit Server VM (build 20.0-b12, mixed mode) ** 3. JAVA_HOME variable: ** COMMAND EXECUTED: echo $JAVA_HOME /usr/lib/jvm/jre-1.6.0-openjdk.x86_64 4. HADOOP version: *** COMMAND EXECUTED: hadoop version Warning: $HADOOP_HOME is deprecated. Hadoop 1.0.4 Subversion https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r 1393290 Compiled by hortonfo on Wed Oct 3 05:13:58 UTC 2012 >From source with checksum fe2baea87c4c81a2c505767f3f9b71f4 Are these still incompatible with eachother? (Hadoop version and java version) Thank you very much. On Tue, Feb 19, 2013 at 10:26 PM, Fatih Haltas wrote: > Thank you all very much > > 19 Şubat 2013 Salı tarihinde Harsh J adlı kullanıcı şöyle yazdı: > >> Oops. I just noticed Hemanth has been answering on a dupe thread as >> well. Lets drop this thread and carry on there :) >> >> On Tue, Feb 19, 2013 at 11:14 PM, Harsh J wrote: >> > Hi, >> > >> > The new error usually happens if you compile using Java 7 and try to >> > run via Java 6 (for example). That is, an incompatibility in the >> > runtimes for the binary artifact produced. >> > >> > On Tue, Feb 19, 2013 at 10:09 PM, Fatih Haltas >> wrote: >> >> Thank you very much Harsh, >> >> >> >> Now, as I promised earlier I am much obliged to you. >> >> >> >> But, now I solved that problem by just changing the directories then >> again >> >> creating a jar file of org. but I am getting this error: >> >> >> >> 1.) What I got >> >> >> -- >> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ hadoop jar flow19028pm.jar >> >> org.myorg.MapReduce /home/hadoop/project/hadoop-data/NetFlow 19_02.out >> >> Warning: $HADOOP_HOME is deprecated. >> >> >> >> Exception in thread "main" java.lang.UnsupportedClassVersionError: >> >> org/myorg/MapReduce : Unsupported major.minor version 51.0 >> >> at java.lang.ClassLoader.defineClass1(Native Method) >> >> at java.lang.ClassLoader.defineClass(ClassLoader.java:634) >> >> at >> >> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >> >> at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) >> >> at java.net.URLClassLoader.access$000(URLClassLoader.java:73) >> >> at java.net.URLClassLoader$1.run(URLClassLoader.java:212) >> >> at java.security.AccessController.doPrivileged(Native Method) >> >> at java.net.URLClassLoader.findClass(URLClassLoader.java:205) >> >> at java.lang.ClassLoader.loadClass(ClassLoader.java:321) >> >> at java.lang.ClassLoader.loadClass(ClassLoader.java:266) >> >> at java.lang.Class.forName0(Native Method) >> >> at java.lang.Class.forName(Class.java:266) >> >> at org.apache.hadoop.util.RunJar.main(RunJar.java:149) >> >> >> >> 2.
Re: Trouble in running MapReduce application
Can you try this ? Pick a class like WordCount from your package and execute this command: javap -classpath -verbose org.myorg.Wordcount | grep version. For e.g. here's what I get for my class: $ javap -verbose WCMapper | grep version minor version: 0 major version: 50 Please paste the output of this - we can verify what the problem is. Thanks Hemanth On Sat, Feb 23, 2013 at 4:45 PM, Fatih Haltas wrote: > Hi again, > > Thanks for your help but now, I am struggling with the same problem on a > machine. As the preivous problem, I just decrease the Java version by Java > 6, but this time I could not solve the problem. > > those are outputs that may explain the situation: > > - > 1. I could not run my own code, to check the system I just tried to run > basic wordcount example without any modification, except package info. > ** > COMMAND EXECUTED: hadoop jar my.jar org.myorg.WordCount NetFlow NetFlow.out > Warning: $HADOOP_HOME is deprecated. > > Exception in thread "main" java.lang.UnsupportedClassVersionError: > org/myorg/WordCount : Unsupported major.minor version 51.0 > at java.lang.ClassLoader.defineClass1(Native Method) > at java.lang.ClassLoader.defineClass(ClassLoader.java:634) > at > java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) > at java.net.URLClassLoader.access$000(URLClassLoader.java:73) > at java.net.URLClassLoader$1.run(URLClassLoader.java:212) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:205) > at java.lang.ClassLoader.loadClass(ClassLoader.java:321) > at java.lang.ClassLoader.loadClass(ClassLoader.java:266) > at java.lang.Class.forName0(Native Method) > at java.lang.Class.forName(Class.java:266) > at org.apache.hadoop.util.RunJar.main(RunJar.java:149) > > ** > 2. Java version: > > COMMAND EXECUTED: java -version > java version "1.6.0_24" > OpenJDK Runtime Environment (IcedTea6 1.11.6) > (rhel-1.33.1.11.6.el5_9-x86_64) > OpenJDK 64-Bit Server VM (build 20.0-b12, mixed mode) > ** > 3. JAVA_HOME variable: > ** > COMMAND EXECUTED: echo $JAVA_HOME > /usr/lib/jvm/jre-1.6.0-openjdk.x86_64 > > 4. HADOOP version: > *** > COMMAND EXECUTED: hadoop version > Warning: $HADOOP_HOME is deprecated. > > Hadoop 1.0.4 > Subversion > https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r > 1393290 > Compiled by hortonfo on Wed Oct 3 05:13:58 UTC 2012 > From source with checksum fe2baea87c4c81a2c505767f3f9b71f4 > > > Are these still incompatible with eachother? (Hadoop version and java > version) > > > Thank you very much. > > > On Tue, Feb 19, 2013 at 10:26 PM, Fatih Haltas wrote: > >> Thank you all very much >> >> 19 Şubat 2013 Salı tarihinde Harsh J adlı kullanıcı şöyle yazdı: >> >>> Oops. I just noticed Hemanth has been answering on a dupe thread as >>> well. Lets drop this thread and carry on there :) >>> >>> On Tue, Feb 19, 2013 at 11:14 PM, Harsh J wrote: >>> > Hi, >>> > >>> > The new error usually happens if you compile using Java 7 and try to >>> > run via Java 6 (for example). That is, an incompatibility in the >>> > runtimes for the binary artifact produced. >>> > >>> > On Tue, Feb 19, 2013 at 10:09 PM, Fatih Haltas >>> wrote: >>> >> Thank you very much Harsh, >>> >> >>> >> Now, as I promised earlier I am much obliged to you. >>> >> >>> >> But, now I solved that problem by just changing the directories then >>> again >>> >> creating a jar file of org. but I am getting this error: >>> >> >>> >> 1.) What I got >>> >> >>> -- >>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ hadoop jar flow19028pm.jar >>> >> org.myorg.MapReduce /home/hadoop/project/hadoop-data/NetFlow 19_02.out >>> >> Warning: $HADOOP_HOME is deprecated. >>> >> >>> >> Exception in thread "main" java.lang.UnsupportedClassVersionError: >>> >> org/myorg/MapReduce : Unsupported major.minor version 51.0 >>> >> at java.lang.ClassLoader.defineClass1(Native Method) >>> >> at java.lang.ClassLoader.defineClass(ClassLoader.java:634) >>> >> at >>> >> >>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >>> >> at >>> java.net.URLClassLoader.defineClass(URLClassLoader.java:277) >>> >> at java.net.URLClassLoader.access
Fwd: Trouble in running MapReduce application
Hi Hemanth; Thanks for your grreat helps, I am really much obliged to you. I solved this problem by changing my java compiler vs. but now though I changed everynodes configuration I am getting this error even I tried to run example of wordcount without making any changes. What may be the reason, I believe that I checked all config files and changed the home variables, also /etc/hosts Here is my problem: [hadoop@ADUAE042-LAP-V logs]$ hadoop jar ../hadoop-examples-1.0.4.jar wordcount /home/hadoop/project/hadoop-data/NetFlow test1353.out Warning: $HADOOP_HOME is deprecated. 13/02/24 13:32:28 INFO input.FileInputFormat: Total input paths to process : 1 13/02/24 13:32:28 INFO util.NativeCodeLoader: Loaded the native-hadoop library 13/02/24 13:32:28 WARN snappy.LoadSnappy: Snappy native library not loaded 13/02/24 13:32:29 INFO mapred.JobClient: Running job: job_201301141457_0034 13/02/24 13:32:30 INFO mapred.JobClient: map 0% reduce 0% 13/02/24 13:32:37 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_m_02_0, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_m_02_0: execvp: No such file or directory 13/02/24 13:32:43 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_r_02_0, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_r_02_0: execvp: No such file or directory 13/02/24 13:32:50 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_m_02_1, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_m_02_1: execvp: No such file or directory 13/02/24 13:32:56 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_r_02_1, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_r_02_1: execvp: No such file or directory 13/02/24 13:33:02 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_m_02_2, Status : FAILED Error initializing attempt_201301141457_0034_m_02_2: java.lang.InternalError at sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769) at sun.misc.URLClassPath.getResource(URLClassPath.java:185) at sun.misc.URLClassPath.getResource(URLClassPath.java:237) at java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113) at java.lang.ClassLoader.getResource(ClassLoader.java:974) at java.lang.ClassLoader.getResource(ClassLoader.java:972) at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075) at java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181) at java.lang.Class.getResourceAsStream(Class.java:2045) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370) at java.security.AccessController.doPrivileged(Native Method) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.(TransformerImpl.java:250) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.(TransformerImpl.java:241) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607) at org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310) at org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559) at org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:416) at org.apache.hadoop.secu
Re: Trouble in running MapReduce application
Hi, Your specifying the input directory in local file system not in HDFS, Copy some text file to using '-put' or '-copyFromLoca'l to HDFS user home directory then try to execute word count by specifying home as input directory. On Sun, Feb 24, 2013 at 3:29 PM, Fatih Haltas wrote: > > > Hi Hemanth; > > Thanks for your grreat helps, > > I am really much obliged to you. > > I solved this problem by changing my java compiler vs. but now though I > changed everynodes configuration I am getting this error even I tried to > run example of wordcount without making any changes. > > What may be the reason, I believe that I checked all config files and > changed the home variables, also /etc/hosts > > Here is my problem: > > [hadoop@ADUAE042-LAP-V logs]$ hadoop jar ../hadoop-examples-1.0.4.jar > wordcount /home/hadoop/project/hadoop-data/NetFlow test1353.out > > Warning: $HADOOP_HOME is deprecated. > > 13/02/24 13:32:28 INFO input.FileInputFormat: Total input paths to process > : 1 > 13/02/24 13:32:28 INFO util.NativeCodeLoader: Loaded the native-hadoop > library > 13/02/24 13:32:28 WARN snappy.LoadSnappy: Snappy native library not loaded > 13/02/24 13:32:29 INFO mapred.JobClient: Running job: job_201301141457_0034 > 13/02/24 13:32:30 INFO mapred.JobClient: map 0% reduce 0% > 13/02/24 13:32:37 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_02_0, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_m_02_0: execvp: No such file or directory > 13/02/24 13:32:43 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_r_02_0, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_r_02_0: execvp: No such file or directory > 13/02/24 13:32:50 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_02_1, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_m_02_1: execvp: No such file or directory > 13/02/24 13:32:56 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_r_02_1, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_r_02_1: execvp: No such file or directory > 13/02/24 13:33:02 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_02_2, Status : FAILED > Error initializing attempt_201301141457_0034_m_02_2: > java.lang.InternalError > at > sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769) > at sun.misc.URLClassPath.getResource(URLClassPath.java:185) > at sun.misc.URLClassPath.getResource(URLClassPath.java:237) > at > java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113) > at java.lang.ClassLoader.getResource(ClassLoader.java:974) > at java.lang.ClassLoader.getResource(ClassLoader.java:972) > at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075) > at > java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181) > at java.lang.Class.getResourceAsStream(Class.java:2045) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370) > at java.security.AccessController.doPrivileged(Native Method) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.(TransformerImpl.java:250) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.(TransformerImpl.java:241) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactory
Re: Trouble in running MapReduce application
Thank you very much but No this is the file in hdfs and it is the exact path of netflow data in hdfs. Hadoop-data is the home hdfs directory, before downgrading my jdk this command worked well 24 Şubat 2013 Pazar tarihinde sudhakara st adlı kullanıcı şöyle yazdı: > Hi, > Your specifying the input directory in local file system not in > HDFS, Copy some text file to using '-put' or '-copyFromLoca'l to HDFS user > home directory then try to execute word count by specifying home as input > directory. > > On Sun, Feb 24, 2013 at 3:29 PM, Fatih Haltas wrote: > > > > Hi Hemanth; > > Thanks for your grreat helps, > > I am really much obliged to you. > > I solved this problem by changing my java compiler vs. but now though I > changed everynodes configuration I am getting this error even I tried to > run example of wordcount without making any changes. > > What may be the reason, I believe that I checked all config files and > changed the home variables, also /etc/hosts > > Here is my problem: > > [hadoop@ADUAE042-LAP-V logs]$ hadoop jar ../hadoop-examples-1.0.4.jar > wordcount /home/hadoop/project/hadoop-data/NetFlow test1353.out > > Warning: $HADOOP_HOME is deprecated. > > 13/02/24 13:32:28 INFO input.FileInputFormat: Total input paths to process > : 1 > 13/02/24 13:32:28 INFO util.NativeCodeLoader: Loaded the native-hadoop > library > 13/02/24 13:32:28 WARN snappy.LoadSnappy: Snappy native library not loaded > 13/02/24 13:32:29 INFO mapred.JobClient: Running job: job_201301141457_0034 > 13/02/24 13:32:30 INFO mapred.JobClient: map 0% reduce 0% > 13/02/24 13:32:37 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_02_0, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_m_02_0: execvp: No such file or directory > 13/02/24 13:32:43 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_r_02_0, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_r_02_0: execvp: No such file or directory > 13/02/24 13:32:50 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_02_1, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_m_02_1: execvp: No such file or directory > 13/02/24 13:32:56 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_r_02_1, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_r_02_1: execvp: No such file or directory > 13/02/24 13:33:02 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_02_2, Status : FAILED > Error initializing attempt_201301141457_0034_m_02_2: > java.lang.InternalError > at > sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769) > at sun.misc.URLClassPath.getResource(URLClassPath.java:185) > at sun.misc.URLClassPath.getResource(URLClassPath.java:237) > at > java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113) > at java.lang.ClassLoader.getResource(ClassLoader.java:974) > at java.lang.ClassLoader.getResource(ClassLoader.java:972) > at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075) > at > java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181) > at java.lang.Class.getResourceAsStream(Class.java:2045) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370) > at java.security.AccessController.doPrivileged(Native M > > -- > > Regards, > . Sudhakara.st > >