Hi,
     Your specifying  the input directory in local file system not in HDFS,
Copy some text file to using '-put' or '-copyFromLoca'l to HDFS user home
directory then try to execute word count by specifying home as input
directory.

On Sun, Feb 24, 2013 at 3:29 PM, Fatih Haltas <fatih.hal...@nyu.edu> wrote:

>
>
> Hi Hemanth;
>
> Thanks for your grreat helps,
>
> I am really much obliged to you.
>
> I solved this problem by changing my java compiler vs. but now though I
> changed everynodes configuration I am getting this error even I tried to
> run example of wordcount without making any changes.
>
> What may be the reason, I believe that I checked all config files and
> changed the home variables, also /etc/hosts
>
> Here is my problem:
> ************************************************************
> [hadoop@ADUAE042-LAP-V logs]$ hadoop jar ../hadoop-examples-1.0.4.jar
> wordcount /home/hadoop/project/hadoop-data/NetFlow test1353.out
>
> Warning: $HADOOP_HOME is deprecated.
>
> 13/02/24 13:32:28 INFO input.FileInputFormat: Total input paths to process
> : 1
> 13/02/24 13:32:28 INFO util.NativeCodeLoader: Loaded the native-hadoop
> library
> 13/02/24 13:32:28 WARN snappy.LoadSnappy: Snappy native library not loaded
> 13/02/24 13:32:29 INFO mapred.JobClient: Running job: job_201301141457_0034
> 13/02/24 13:32:30 INFO mapred.JobClient:  map 0% reduce 0%
> 13/02/24 13:32:37 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_m_000002_0, Status : FAILED
> java.lang.Throwable: Child Error
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
> Caused by: java.io.IOException: Task process exit with nonzero status of 1.
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
>
> attempt_201301141457_0034_m_000002_0: execvp: No such file or directory
> 13/02/24 13:32:43 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_r_000002_0, Status : FAILED
> java.lang.Throwable: Child Error
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
> Caused by: java.io.IOException: Task process exit with nonzero status of 1.
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
>
> attempt_201301141457_0034_r_000002_0: execvp: No such file or directory
> 13/02/24 13:32:50 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_m_000002_1, Status : FAILED
> java.lang.Throwable: Child Error
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
> Caused by: java.io.IOException: Task process exit with nonzero status of 1.
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
>
> attempt_201301141457_0034_m_000002_1: execvp: No such file or directory
> 13/02/24 13:32:56 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_r_000002_1, Status : FAILED
> java.lang.Throwable: Child Error
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
> Caused by: java.io.IOException: Task process exit with nonzero status of 1.
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
>
> attempt_201301141457_0034_r_000002_1: execvp: No such file or directory
> 13/02/24 13:33:02 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_m_000002_2, Status : FAILED
> Error initializing attempt_201301141457_0034_m_000002_2:
> java.lang.InternalError
>         at
> sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769)
>         at sun.misc.URLClassPath.getResource(URLClassPath.java:185)
>         at sun.misc.URLClassPath.getResource(URLClassPath.java:237)
>         at
> java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113)
>         at java.lang.ClassLoader.getResource(ClassLoader.java:974)
>         at java.lang.ClassLoader.getResource(ClassLoader.java:972)
>         at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075)
>         at
> java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181)
>         at java.lang.Class.getResourceAsStream(Class.java:2045)
>         at
> com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at
> com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366)
>         at
> com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:250)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:241)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607)
>         at
> org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310)
>         at
> org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559)
>         at
> org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:416)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>         at
> org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1203)
>         at
> org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1118)
>         at
> org.apache.hadoop.mapred.TaskTracker$5.run(TaskTracker.java:2430)
>         at java.lang.Thread.run(Thread.java:679)
> Caused by: java.io.FileNotFoundException:
> /usr/lib/jvm/java-1.6.0-openjdk-1.6.0.0.x86_64/jre/lib/resources.jar
>         at
> sun.misc.URLClassPath$JarLoader.getJarFile(URLClassPath.java:658)
>         at
> sun.misc.URLClassPath$JarLoader.access$600(URLClassPath.java:555)
>         at sun.misc.URLClassPath$JarLoader$1.run(URLClassPath.java:622)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at
> sun.misc.URLClassPath$JarLoader.ensureOpen(URLClassPath.java:614)
>         at
> sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:767)
>         ... 27 more
>
>  13/02/24 13:33:02 WARN mapred.JobClient: Error reading task
> outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000002_2&filter=stdout
> 13/02/24 13:33:02 WARN mapred.JobClient: Error reading task
> outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000002_2&filter=stderr
> 13/02/24 13:33:08 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_m_000001_0, Status : FAILED
> Error initializing attempt_201301141457_0034_m_000001_0:
> java.lang.InternalError
>         at
> sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769)
>         at sun.misc.URLClassPath.getResource(URLClassPath.java:185)
>         at sun.misc.URLClassPath.getResource(URLClassPath.java:237)
>         at
> java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113)
>         at java.lang.ClassLoader.getResource(ClassLoader.java:974)
>         at java.lang.ClassLoader.getResource(ClassLoader.java:972)
>         at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075)
>         at
> java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181)
>         at java.lang.Class.getResourceAsStream(Class.java:2045)
>         at
> com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at
> com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366)
>         at
> com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:250)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:241)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607)
>         at
> org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310)
>         at
> org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559)
>         at
> org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:416)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>         at
> org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1203)
>         at
> org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1118)
>         at
> org.apache.hadoop.mapred.TaskTracker$5.run(TaskTracker.java:2430)
>         at java.lang.Thread.run(Thread.java:679)
> Caused by: java.io.FileNotFoundException:
> /usr/lib/jvm/java-1.6.0-openjdk-1.6.0.0.x86_64/jre/lib/resources.jar
>         at
> sun.misc.URLClassPath$JarLoader.getJarFile(URLClassPath.java:658)
>         at
> sun.misc.URLClassPath$JarLoader.access$600(URLClassPath.java:555)
>         at sun.misc.URLClassPath$JarLoader$1.run(URLClassPath.java:622)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at
> sun.misc.URLClassPath$JarLoader.ensureOpen(URLClassPath.java:614)
>         at
> sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:767)
>         ... 27 more
>
>  13/02/24 13:33:08 WARN mapred.JobClient: Error reading task
> outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000001_0&filter=stdout
> 13/02/24 13:33:08 WARN mapred.JobClient: Error reading task
> outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000001_0&filter=stderr
> 13/02/24 13:33:11 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_r_000001_0, Status : FAILED
> Error initializing attempt_201301141457_0034_r_000001_0:
> java.lang.InternalError
>         at
> sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769)
>         at sun.misc.URLClassPath.getResource(URLClassPath.java:185)
>         at sun.misc.URLClassPath.getResource(URLClassPath.java:237)
>         at
> java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113)
>         at java.lang.ClassLoader.getResource(ClassLoader.java:974)
>         at java.lang.ClassLoader.getResource(ClassLoader.java:972)
>         at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075)
>         at
> java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181)
>         at java.lang.Class.getResourceAsStream(Class.java:2045)
>         at
> com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at
> com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366)
>         at
> com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:250)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:241)
>         at
> com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607)
>         at
> org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310)
>         at
> org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559)
>         at
> org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:416)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>         at
> org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1203)
>         at
> org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1118)
>         at
> org.apache.hadoop.mapred.TaskTracker$5.run(TaskTracker.java:2430)
>         at java.lang.Thread.run(Thread.java:679)
> Caused by: java.io.FileNotFoundException:
> /usr/lib/jvm/java-1.6.0-openjdk-1.6.0.0.x86_64/jre/lib/resources.jar
>         at
> sun.misc.URLClassPath$JarLoader.getJarFile(URLClassPath.java:658)
>         at
> sun.misc.URLClassPath$JarLoader.access$600(URLClassPath.java:555)
>         at sun.misc.URLClassPath$JarLoader$1.run(URLClassPath.java:622)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at
> sun.misc.URLClassPath$JarLoader.ensureOpen(URLClassPath.java:614)
>         at
> sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:767)
>         ... 27 more
>
>  13/02/24 13:33:11 WARN mapred.JobClient: Error reading task
> outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_r_000001_0&filter=stdout
> 13/02/24 13:33:11 WARN mapred.JobClient: Error reading task
> outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_r_000001_0&filter=stderr
> 13/02/24 13:33:19 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_m_000001_1, Status : FAILED
> java.lang.Throwable: Child Error
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
> Caused by: java.io.IOException: Task process exit with nonzero status of 1.
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
>
> attempt_201301141457_0034_m_000001_1: execvp: No such file or directory
> 13/02/24 13:33:25 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_r_000001_1, Status : FAILED
> java.lang.Throwable: Child Error
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
> Caused by: java.io.IOException: Task process exit with nonzero status of 1.
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
>
> attempt_201301141457_0034_r_000001_1: execvp: No such file or directory
> 13/02/24 13:33:31 INFO mapred.JobClient: Task Id :
> attempt_201301141457_0034_m_000001_2, Status : FAILED
> java.lang.Throwable: Child Error
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
> Caused by: java.io.IOException: Task process exit with nonzero status of 1.
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
>
> attempt_201301141457_0034_m_000001_2: execvp: No such file or directory
> 13/02/24 13:33:37 INFO mapred.JobClient: Job complete:
> job_201301141457_0034
> 13/02/24 13:33:37 INFO mapred.JobClient: Counters: 4
> 13/02/24 13:33:37 INFO mapred.JobClient:   Job Counters
> 13/02/24 13:33:37 INFO mapred.JobClient:     SLOTS_MILLIS_MAPS=15112
> 13/02/24 13:33:37 INFO mapred.JobClient:     Total time spent by all
> reduces waiting after reserving slots (ms)=0
> 13/02/24 13:33:37 INFO mapred.JobClient:     Total time spent by all maps
> waiting after reserving slots (ms)=0
> 13/02/24 13:33:37 INFO mapred.JobClient:     SLOTS_MILLIS_REDUCES=9091
>
>
>
> On Sat, Feb 23, 2013 at 5:05 PM, Hemanth Yamijala <
> yhema...@thoughtworks.com> wrote:
>
>> Can you try this ? Pick a class like WordCount from your package and
>> execute this command:
>>
>> javap -classpath <path to your jar> -verbose org.myorg.Wordcount | grep
>> version.
>>
>> For e.g. here's what I get for my class:
>>
>> $ javap -verbose WCMapper | grep version
>>   minor version: 0
>>   major version: 50
>>
>> Please paste the output of this - we can verify what the problem is.
>>
>> Thanks
>> Hemanth
>>
>>
>> On Sat, Feb 23, 2013 at 4:45 PM, Fatih Haltas <fatih.hal...@nyu.edu>wrote:
>>
>>> Hi again,
>>>
>>> Thanks for your help but now, I am struggling with the same problem on a
>>> machine. As the preivous problem, I just decrease the Java version by Java
>>> 6, but this time I could not solve the problem.
>>>
>>> those are outputs that may explain the situation:
>>>
>>> ---------------------------------------------------------------------------------------------------------------------------------------------
>>> 1. I could not run my own code, to check the system I just tried to run
>>> basic wordcount example without any modification, except package info.
>>> **************************************************
>>> COMMAND EXECUTED: hadoop jar my.jar org.myorg.WordCount NetFlow
>>> NetFlow.out
>>> Warning: $HADOOP_HOME is deprecated.
>>>
>>> Exception in thread "main" java.lang.UnsupportedClassVersionError:
>>> org/myorg/WordCount : Unsupported major.minor version 51.0
>>>          at java.lang.ClassLoader.defineClass1(Native Method)
>>>         at java.lang.ClassLoader.defineClass(ClassLoader.java:634)
>>>         at
>>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
>>>         at java.net.URLClassLoader.defineClass(URLClassLoader.java:277)
>>>         at java.net.URLClassLoader.access$000(URLClassLoader.java:73)
>>>         at java.net.URLClassLoader$1.run(URLClassLoader.java:212)
>>>         at java.security.AccessController.doPrivileged(Native Method)
>>>         at java.net.URLClassLoader.findClass(URLClassLoader.java:205)
>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:321)
>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:266)
>>>         at java.lang.Class.forName0(Native Method)
>>>         at java.lang.Class.forName(Class.java:266)
>>>         at org.apache.hadoop.util.RunJar.main(RunJar.java:149)
>>>
>>> **************************************************************************************
>>> 2. Java version:
>>> ********************************
>>> COMMAND EXECUTED: java -version
>>> java version "1.6.0_24"
>>> OpenJDK Runtime Environment (IcedTea6 1.11.6)
>>> (rhel-1.33.1.11.6.el5_9-x86_64)
>>> OpenJDK 64-Bit Server VM (build 20.0-b12, mixed mode)
>>> **********************************
>>> 3. JAVA_HOME variable:
>>> **********************************
>>> COMMAND EXECUTED: echo $JAVA_HOME
>>> /usr/lib/jvm/jre-1.6.0-openjdk.x86_64
>>> ********************************************
>>> 4. HADOOP version:
>>> *******************************************
>>> COMMAND EXECUTED: hadoop version
>>> Warning: $HADOOP_HOME is deprecated.
>>>
>>> Hadoop 1.0.4
>>> Subversion
>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
>>> 1393290
>>> Compiled by hortonfo on Wed Oct  3 05:13:58 UTC 2012
>>> From source with checksum fe2baea87c4c81a2c505767f3f9b71f4
>>> ********************************************************
>>>
>>> Are these still incompatible with eachother? (Hadoop version and java
>>> version)
>>>
>>>
>>> Thank you very much.
>>>
>>>
>>> On Tue, Feb 19, 2013 at 10:26 PM, Fatih Haltas <fatih.hal...@nyu.edu>wrote:
>>>
>>>> Thank you all very much
>>>>
>>>> 19 Şubat 2013 Salı tarihinde Harsh J adlı kullanıcı şöyle yazdı:
>>>>
>>>>> Oops. I just noticed Hemanth has been answering on a dupe thread as
>>>>> well. Lets drop this thread and carry on there :)
>>>>>
>>>>> On Tue, Feb 19, 2013 at 11:14 PM, Harsh J <ha...@cloudera.com> wrote:
>>>>> > Hi,
>>>>> >
>>>>> > The new error usually happens if you compile using Java 7 and try to
>>>>> > run via Java 6 (for example). That is, an incompatibility in the
>>>>> > runtimes for the binary artifact produced.
>>>>> >
>>>>> > On Tue, Feb 19, 2013 at 10:09 PM, Fatih Haltas <fatih.hal...@nyu.edu>
>>>>> wrote:
>>>>> >> Thank you very much Harsh,
>>>>> >>
>>>>> >> Now, as I promised earlier I am much obliged to you.
>>>>> >>
>>>>> >> But, now I solved that problem by just changing the directories
>>>>> then again
>>>>> >> creating a jar file of org. but I am getting this error:
>>>>> >>
>>>>> >> 1.) What I got
>>>>> >>
>>>>> ------------------------------------------------------------------------------
>>>>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ hadoop jar
>>>>> flow19028pm.jar
>>>>> >> org.myorg.MapReduce /home/hadoop/project/hadoop-data/NetFlow
>>>>> 19_02.out
>>>>> >> Warning: $HADOOP_HOME is deprecated.
>>>>> >>
>>>>> >> Exception in thread "main" java.lang.UnsupportedClassVersionError:
>>>>> >> org/myorg/MapReduce : Unsupported major.minor version 51.0
>>>>> >>         at java.lang.ClassLoader.defineClass1(Native Method)
>>>>> >>         at java.lang.ClassLoader.defineClass(ClassLoader.java:634)
>>>>> >>         at
>>>>> >>
>>>>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
>>>>> >>         at
>>>>> java.net.URLClassLoader.defineClass(URLClassLoader.java:277)
>>>>> >>         at
>>>>> java.net.URLClassLoader.access$000(URLClassLoader.java:73)
>>>>> >>         at java.net.URLClassLoader$1.run(URLClassLoader.java:212)
>>>>> >>         at java.security.AccessController.doPrivileged(Native
>>>>> Method)
>>>>> >>         at
>>>>> java.net.URLClassLoader.findClass(URLClassLoader.java:205)
>>>>> >>         at java.lang.ClassLoader.loadClass(ClassLoader.java:321)
>>>>> >>         at java.lang.ClassLoader.loadClass(ClassLoader.java:266)
>>>>> >>         at java.lang.Class.forName0(Native Method)
>>>>> >>         at java.lang.Class.forName(Class.java:266)
>>>>> >>         at org.apache.hadoop.util.RunJar.main(RunJar.java:149)
>>>>> >>
>>>>> >> 2.) How I create my jar
>>>>> >>
>>>>> -------------------------------------------------------------------------------------
>>>>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar cvf flow19028pm.jar
>>>>> org
>>>>> >> added manifest
>>>>> >> adding: org/(in = 0) (out= 0)(stored 0%)
>>>>> >> adding: org/myorg/(in = 0) (out= 0)(stored 0%)
>>>>> >> adding: org/myorg/MapReduce$FlowPortReducer.class(in = 1661) (out=
>>>>> >> 690)(deflated 58%)
>>>>> >> adding: org/myorg/MapReduce.class(in = 1587) (out= 903)(deflated
>>>>> 43%)
>>>>> >> adding: org/myorg/MapReduce$FlowPortMapper.class(in = 1874) (out=
>>>>> >> 823)(deflated 56%)
>>>>> >>
>>>>> >> 3.) Content of my jar file
>>>>> >>
>>>>> ---------------------------------------------------------------------------------------
>>>>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar tf flow19028pm.jar
>>>>> >> META-INF/
>>>>> >> META-INF/MANIFEST.MF
>>>>> >> org/
>>>>> >> org/myorg/
>>>>> >> org/myorg/MapReduce$FlowPortReducer.class
>>>>> >> org/myorg/MapReduce.class
>>>>> >> org/myorg/MapReduce$FlowPortMapper.class
>>>>> >>
>>>>> -----------------------------------------------------------------------------------------
>>>>> >>
>>>>> >>
>>>>> >> Thank you very much.
>>>>> >>
>>>>> >>
>>>>> >> On Tue, Feb 19, 2013 at 8:20 PM, Harsh J <ha...@cloudera.com>
>>>>> wrote:
>>>>> >>>
>>>>> >>> Your point (4) explains the problem. The jar packed structure
>>>>> should
>>>>> >>> look like the below, and not how it is presently (one extra top
>>>>> level
>>>>> >>> dir is present):
>>>>> >>>
>>>>> >>> META-INF/
>>>>> >>> META-INF/MANIFEST.MF
>>>>> >>> org/
>>>>> >>> org/myorg/
>>>>> >>> org/myorg/WordCount.class
>>>>> >>> org/myorg/WordCount$TokenizerMapper.class
>>>>> >>> org/myorg/WordCount$IntSumReducer.clas--
>>>>> Harsh J
>>>>>
>>>>
>>>
>>
>
>


-- 

Regards,
.....  Sudhakara.st

Reply via email to