Hi, Your specifying the input directory in local file system not in HDFS, Copy some text file to using '-put' or '-copyFromLoca'l to HDFS user home directory then try to execute word count by specifying home as input directory.
On Sun, Feb 24, 2013 at 3:29 PM, Fatih Haltas <fatih.hal...@nyu.edu> wrote: > > > Hi Hemanth; > > Thanks for your grreat helps, > > I am really much obliged to you. > > I solved this problem by changing my java compiler vs. but now though I > changed everynodes configuration I am getting this error even I tried to > run example of wordcount without making any changes. > > What may be the reason, I believe that I checked all config files and > changed the home variables, also /etc/hosts > > Here is my problem: > ************************************************************ > [hadoop@ADUAE042-LAP-V logs]$ hadoop jar ../hadoop-examples-1.0.4.jar > wordcount /home/hadoop/project/hadoop-data/NetFlow test1353.out > > Warning: $HADOOP_HOME is deprecated. > > 13/02/24 13:32:28 INFO input.FileInputFormat: Total input paths to process > : 1 > 13/02/24 13:32:28 INFO util.NativeCodeLoader: Loaded the native-hadoop > library > 13/02/24 13:32:28 WARN snappy.LoadSnappy: Snappy native library not loaded > 13/02/24 13:32:29 INFO mapred.JobClient: Running job: job_201301141457_0034 > 13/02/24 13:32:30 INFO mapred.JobClient: map 0% reduce 0% > 13/02/24 13:32:37 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_000002_0, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_m_000002_0: execvp: No such file or directory > 13/02/24 13:32:43 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_r_000002_0, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_r_000002_0: execvp: No such file or directory > 13/02/24 13:32:50 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_000002_1, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_m_000002_1: execvp: No such file or directory > 13/02/24 13:32:56 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_r_000002_1, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_r_000002_1: execvp: No such file or directory > 13/02/24 13:33:02 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_000002_2, Status : FAILED > Error initializing attempt_201301141457_0034_m_000002_2: > java.lang.InternalError > at > sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769) > at sun.misc.URLClassPath.getResource(URLClassPath.java:185) > at sun.misc.URLClassPath.getResource(URLClassPath.java:237) > at > java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113) > at java.lang.ClassLoader.getResource(ClassLoader.java:974) > at java.lang.ClassLoader.getResource(ClassLoader.java:972) > at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075) > at > java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181) > at java.lang.Class.getResourceAsStream(Class.java:2045) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370) > at java.security.AccessController.doPrivileged(Native Method) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:250) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:241) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607) > at > org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310) > at > org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559) > at > org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:416) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121) > at > org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1203) > at > org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1118) > at > org.apache.hadoop.mapred.TaskTracker$5.run(TaskTracker.java:2430) > at java.lang.Thread.run(Thread.java:679) > Caused by: java.io.FileNotFoundException: > /usr/lib/jvm/java-1.6.0-openjdk-1.6.0.0.x86_64/jre/lib/resources.jar > at > sun.misc.URLClassPath$JarLoader.getJarFile(URLClassPath.java:658) > at > sun.misc.URLClassPath$JarLoader.access$600(URLClassPath.java:555) > at sun.misc.URLClassPath$JarLoader$1.run(URLClassPath.java:622) > at java.security.AccessController.doPrivileged(Native Method) > at > sun.misc.URLClassPath$JarLoader.ensureOpen(URLClassPath.java:614) > at > sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:767) > ... 27 more > > 13/02/24 13:33:02 WARN mapred.JobClient: Error reading task > outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000002_2&filter=stdout > 13/02/24 13:33:02 WARN mapred.JobClient: Error reading task > outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000002_2&filter=stderr > 13/02/24 13:33:08 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_000001_0, Status : FAILED > Error initializing attempt_201301141457_0034_m_000001_0: > java.lang.InternalError > at > sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769) > at sun.misc.URLClassPath.getResource(URLClassPath.java:185) > at sun.misc.URLClassPath.getResource(URLClassPath.java:237) > at > java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113) > at java.lang.ClassLoader.getResource(ClassLoader.java:974) > at java.lang.ClassLoader.getResource(ClassLoader.java:972) > at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075) > at > java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181) > at java.lang.Class.getResourceAsStream(Class.java:2045) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370) > at java.security.AccessController.doPrivileged(Native Method) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:250) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:241) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607) > at > org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310) > at > org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559) > at > org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:416) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121) > at > org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1203) > at > org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1118) > at > org.apache.hadoop.mapred.TaskTracker$5.run(TaskTracker.java:2430) > at java.lang.Thread.run(Thread.java:679) > Caused by: java.io.FileNotFoundException: > /usr/lib/jvm/java-1.6.0-openjdk-1.6.0.0.x86_64/jre/lib/resources.jar > at > sun.misc.URLClassPath$JarLoader.getJarFile(URLClassPath.java:658) > at > sun.misc.URLClassPath$JarLoader.access$600(URLClassPath.java:555) > at sun.misc.URLClassPath$JarLoader$1.run(URLClassPath.java:622) > at java.security.AccessController.doPrivileged(Native Method) > at > sun.misc.URLClassPath$JarLoader.ensureOpen(URLClassPath.java:614) > at > sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:767) > ... 27 more > > 13/02/24 13:33:08 WARN mapred.JobClient: Error reading task > outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000001_0&filter=stdout > 13/02/24 13:33:08 WARN mapred.JobClient: Error reading task > outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000001_0&filter=stderr > 13/02/24 13:33:11 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_r_000001_0, Status : FAILED > Error initializing attempt_201301141457_0034_r_000001_0: > java.lang.InternalError > at > sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769) > at sun.misc.URLClassPath.getResource(URLClassPath.java:185) > at sun.misc.URLClassPath.getResource(URLClassPath.java:237) > at > java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113) > at java.lang.ClassLoader.getResource(ClassLoader.java:974) > at java.lang.ClassLoader.getResource(ClassLoader.java:972) > at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075) > at > java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181) > at java.lang.Class.getResourceAsStream(Class.java:2045) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370) > at java.security.AccessController.doPrivileged(Native Method) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366) > at > com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:250) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:241) > at > com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607) > at > org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310) > at > org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559) > at > org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:416) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121) > at > org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1203) > at > org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1118) > at > org.apache.hadoop.mapred.TaskTracker$5.run(TaskTracker.java:2430) > at java.lang.Thread.run(Thread.java:679) > Caused by: java.io.FileNotFoundException: > /usr/lib/jvm/java-1.6.0-openjdk-1.6.0.0.x86_64/jre/lib/resources.jar > at > sun.misc.URLClassPath$JarLoader.getJarFile(URLClassPath.java:658) > at > sun.misc.URLClassPath$JarLoader.access$600(URLClassPath.java:555) > at sun.misc.URLClassPath$JarLoader$1.run(URLClassPath.java:622) > at java.security.AccessController.doPrivileged(Native Method) > at > sun.misc.URLClassPath$JarLoader.ensureOpen(URLClassPath.java:614) > at > sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:767) > ... 27 more > > 13/02/24 13:33:11 WARN mapred.JobClient: Error reading task > outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_r_000001_0&filter=stdout > 13/02/24 13:33:11 WARN mapred.JobClient: Error reading task > outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_r_000001_0&filter=stderr > 13/02/24 13:33:19 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_000001_1, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_m_000001_1: execvp: No such file or directory > 13/02/24 13:33:25 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_r_000001_1, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_r_000001_1: execvp: No such file or directory > 13/02/24 13:33:31 INFO mapred.JobClient: Task Id : > attempt_201301141457_0034_m_000001_2, Status : FAILED > java.lang.Throwable: Child Error > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) > Caused by: java.io.IOException: Task process exit with nonzero status of 1. > at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) > > attempt_201301141457_0034_m_000001_2: execvp: No such file or directory > 13/02/24 13:33:37 INFO mapred.JobClient: Job complete: > job_201301141457_0034 > 13/02/24 13:33:37 INFO mapred.JobClient: Counters: 4 > 13/02/24 13:33:37 INFO mapred.JobClient: Job Counters > 13/02/24 13:33:37 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=15112 > 13/02/24 13:33:37 INFO mapred.JobClient: Total time spent by all > reduces waiting after reserving slots (ms)=0 > 13/02/24 13:33:37 INFO mapred.JobClient: Total time spent by all maps > waiting after reserving slots (ms)=0 > 13/02/24 13:33:37 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=9091 > > > > On Sat, Feb 23, 2013 at 5:05 PM, Hemanth Yamijala < > yhema...@thoughtworks.com> wrote: > >> Can you try this ? Pick a class like WordCount from your package and >> execute this command: >> >> javap -classpath <path to your jar> -verbose org.myorg.Wordcount | grep >> version. >> >> For e.g. here's what I get for my class: >> >> $ javap -verbose WCMapper | grep version >> minor version: 0 >> major version: 50 >> >> Please paste the output of this - we can verify what the problem is. >> >> Thanks >> Hemanth >> >> >> On Sat, Feb 23, 2013 at 4:45 PM, Fatih Haltas <fatih.hal...@nyu.edu>wrote: >> >>> Hi again, >>> >>> Thanks for your help but now, I am struggling with the same problem on a >>> machine. As the preivous problem, I just decrease the Java version by Java >>> 6, but this time I could not solve the problem. >>> >>> those are outputs that may explain the situation: >>> >>> --------------------------------------------------------------------------------------------------------------------------------------------- >>> 1. I could not run my own code, to check the system I just tried to run >>> basic wordcount example without any modification, except package info. >>> ************************************************** >>> COMMAND EXECUTED: hadoop jar my.jar org.myorg.WordCount NetFlow >>> NetFlow.out >>> Warning: $HADOOP_HOME is deprecated. >>> >>> Exception in thread "main" java.lang.UnsupportedClassVersionError: >>> org/myorg/WordCount : Unsupported major.minor version 51.0 >>> at java.lang.ClassLoader.defineClass1(Native Method) >>> at java.lang.ClassLoader.defineClass(ClassLoader.java:634) >>> at >>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >>> at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) >>> at java.net.URLClassLoader.access$000(URLClassLoader.java:73) >>> at java.net.URLClassLoader$1.run(URLClassLoader.java:212) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at java.net.URLClassLoader.findClass(URLClassLoader.java:205) >>> at java.lang.ClassLoader.loadClass(ClassLoader.java:321) >>> at java.lang.ClassLoader.loadClass(ClassLoader.java:266) >>> at java.lang.Class.forName0(Native Method) >>> at java.lang.Class.forName(Class.java:266) >>> at org.apache.hadoop.util.RunJar.main(RunJar.java:149) >>> >>> ************************************************************************************** >>> 2. Java version: >>> ******************************** >>> COMMAND EXECUTED: java -version >>> java version "1.6.0_24" >>> OpenJDK Runtime Environment (IcedTea6 1.11.6) >>> (rhel-1.33.1.11.6.el5_9-x86_64) >>> OpenJDK 64-Bit Server VM (build 20.0-b12, mixed mode) >>> ********************************** >>> 3. JAVA_HOME variable: >>> ********************************** >>> COMMAND EXECUTED: echo $JAVA_HOME >>> /usr/lib/jvm/jre-1.6.0-openjdk.x86_64 >>> ******************************************** >>> 4. HADOOP version: >>> ******************************************* >>> COMMAND EXECUTED: hadoop version >>> Warning: $HADOOP_HOME is deprecated. >>> >>> Hadoop 1.0.4 >>> Subversion >>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r >>> 1393290 >>> Compiled by hortonfo on Wed Oct 3 05:13:58 UTC 2012 >>> From source with checksum fe2baea87c4c81a2c505767f3f9b71f4 >>> ******************************************************** >>> >>> Are these still incompatible with eachother? (Hadoop version and java >>> version) >>> >>> >>> Thank you very much. >>> >>> >>> On Tue, Feb 19, 2013 at 10:26 PM, Fatih Haltas <fatih.hal...@nyu.edu>wrote: >>> >>>> Thank you all very much >>>> >>>> 19 Şubat 2013 Salı tarihinde Harsh J adlı kullanıcı şöyle yazdı: >>>> >>>>> Oops. I just noticed Hemanth has been answering on a dupe thread as >>>>> well. Lets drop this thread and carry on there :) >>>>> >>>>> On Tue, Feb 19, 2013 at 11:14 PM, Harsh J <ha...@cloudera.com> wrote: >>>>> > Hi, >>>>> > >>>>> > The new error usually happens if you compile using Java 7 and try to >>>>> > run via Java 6 (for example). That is, an incompatibility in the >>>>> > runtimes for the binary artifact produced. >>>>> > >>>>> > On Tue, Feb 19, 2013 at 10:09 PM, Fatih Haltas <fatih.hal...@nyu.edu> >>>>> wrote: >>>>> >> Thank you very much Harsh, >>>>> >> >>>>> >> Now, as I promised earlier I am much obliged to you. >>>>> >> >>>>> >> But, now I solved that problem by just changing the directories >>>>> then again >>>>> >> creating a jar file of org. but I am getting this error: >>>>> >> >>>>> >> 1.) What I got >>>>> >> >>>>> ------------------------------------------------------------------------------ >>>>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ hadoop jar >>>>> flow19028pm.jar >>>>> >> org.myorg.MapReduce /home/hadoop/project/hadoop-data/NetFlow >>>>> 19_02.out >>>>> >> Warning: $HADOOP_HOME is deprecated. >>>>> >> >>>>> >> Exception in thread "main" java.lang.UnsupportedClassVersionError: >>>>> >> org/myorg/MapReduce : Unsupported major.minor version 51.0 >>>>> >> at java.lang.ClassLoader.defineClass1(Native Method) >>>>> >> at java.lang.ClassLoader.defineClass(ClassLoader.java:634) >>>>> >> at >>>>> >> >>>>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >>>>> >> at >>>>> java.net.URLClassLoader.defineClass(URLClassLoader.java:277) >>>>> >> at >>>>> java.net.URLClassLoader.access$000(URLClassLoader.java:73) >>>>> >> at java.net.URLClassLoader$1.run(URLClassLoader.java:212) >>>>> >> at java.security.AccessController.doPrivileged(Native >>>>> Method) >>>>> >> at >>>>> java.net.URLClassLoader.findClass(URLClassLoader.java:205) >>>>> >> at java.lang.ClassLoader.loadClass(ClassLoader.java:321) >>>>> >> at java.lang.ClassLoader.loadClass(ClassLoader.java:266) >>>>> >> at java.lang.Class.forName0(Native Method) >>>>> >> at java.lang.Class.forName(Class.java:266) >>>>> >> at org.apache.hadoop.util.RunJar.main(RunJar.java:149) >>>>> >> >>>>> >> 2.) How I create my jar >>>>> >> >>>>> ------------------------------------------------------------------------------------- >>>>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar cvf flow19028pm.jar >>>>> org >>>>> >> added manifest >>>>> >> adding: org/(in = 0) (out= 0)(stored 0%) >>>>> >> adding: org/myorg/(in = 0) (out= 0)(stored 0%) >>>>> >> adding: org/myorg/MapReduce$FlowPortReducer.class(in = 1661) (out= >>>>> >> 690)(deflated 58%) >>>>> >> adding: org/myorg/MapReduce.class(in = 1587) (out= 903)(deflated >>>>> 43%) >>>>> >> adding: org/myorg/MapReduce$FlowPortMapper.class(in = 1874) (out= >>>>> >> 823)(deflated 56%) >>>>> >> >>>>> >> 3.) Content of my jar file >>>>> >> >>>>> --------------------------------------------------------------------------------------- >>>>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar tf flow19028pm.jar >>>>> >> META-INF/ >>>>> >> META-INF/MANIFEST.MF >>>>> >> org/ >>>>> >> org/myorg/ >>>>> >> org/myorg/MapReduce$FlowPortReducer.class >>>>> >> org/myorg/MapReduce.class >>>>> >> org/myorg/MapReduce$FlowPortMapper.class >>>>> >> >>>>> ----------------------------------------------------------------------------------------- >>>>> >> >>>>> >> >>>>> >> Thank you very much. >>>>> >> >>>>> >> >>>>> >> On Tue, Feb 19, 2013 at 8:20 PM, Harsh J <ha...@cloudera.com> >>>>> wrote: >>>>> >>> >>>>> >>> Your point (4) explains the problem. The jar packed structure >>>>> should >>>>> >>> look like the below, and not how it is presently (one extra top >>>>> level >>>>> >>> dir is present): >>>>> >>> >>>>> >>> META-INF/ >>>>> >>> META-INF/MANIFEST.MF >>>>> >>> org/ >>>>> >>> org/myorg/ >>>>> >>> org/myorg/WordCount.class >>>>> >>> org/myorg/WordCount$TokenizerMapper.class >>>>> >>> org/myorg/WordCount$IntSumReducer.clas-- >>>>> Harsh J >>>>> >>>> >>> >> > > -- Regards, ..... Sudhakara.st