Hi Hemanth; Thanks for your grreat helps,
I am really much obliged to you. I solved this problem by changing my java compiler vs. but now though I changed everynodes configuration I am getting this error even I tried to run example of wordcount without making any changes. What may be the reason, I believe that I checked all config files and changed the home variables, also /etc/hosts Here is my problem: ************************************************************ [hadoop@ADUAE042-LAP-V logs]$ hadoop jar ../hadoop-examples-1.0.4.jar wordcount /home/hadoop/project/hadoop-data/NetFlow test1353.out Warning: $HADOOP_HOME is deprecated. 13/02/24 13:32:28 INFO input.FileInputFormat: Total input paths to process : 1 13/02/24 13:32:28 INFO util.NativeCodeLoader: Loaded the native-hadoop library 13/02/24 13:32:28 WARN snappy.LoadSnappy: Snappy native library not loaded 13/02/24 13:32:29 INFO mapred.JobClient: Running job: job_201301141457_0034 13/02/24 13:32:30 INFO mapred.JobClient: map 0% reduce 0% 13/02/24 13:32:37 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_m_000002_0, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_m_000002_0: execvp: No such file or directory 13/02/24 13:32:43 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_r_000002_0, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_r_000002_0: execvp: No such file or directory 13/02/24 13:32:50 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_m_000002_1, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_m_000002_1: execvp: No such file or directory 13/02/24 13:32:56 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_r_000002_1, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_r_000002_1: execvp: No such file or directory 13/02/24 13:33:02 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_m_000002_2, Status : FAILED Error initializing attempt_201301141457_0034_m_000002_2: java.lang.InternalError at sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769) at sun.misc.URLClassPath.getResource(URLClassPath.java:185) at sun.misc.URLClassPath.getResource(URLClassPath.java:237) at java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113) at java.lang.ClassLoader.getResource(ClassLoader.java:974) at java.lang.ClassLoader.getResource(ClassLoader.java:972) at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075) at java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181) at java.lang.Class.getResourceAsStream(Class.java:2045) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370) at java.security.AccessController.doPrivileged(Native Method) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:250) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:241) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607) at org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310) at org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559) at org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:416) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121) at org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1203) at org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1118) at org.apache.hadoop.mapred.TaskTracker$5.run(TaskTracker.java:2430) at java.lang.Thread.run(Thread.java:679) Caused by: java.io.FileNotFoundException: /usr/lib/jvm/java-1.6.0-openjdk-1.6.0.0.x86_64/jre/lib/resources.jar at sun.misc.URLClassPath$JarLoader.getJarFile(URLClassPath.java:658) at sun.misc.URLClassPath$JarLoader.access$600(URLClassPath.java:555) at sun.misc.URLClassPath$JarLoader$1.run(URLClassPath.java:622) at java.security.AccessController.doPrivileged(Native Method) at sun.misc.URLClassPath$JarLoader.ensureOpen(URLClassPath.java:614) at sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:767) ... 27 more 13/02/24 13:33:02 WARN mapred.JobClient: Error reading task outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000002_2&filter=stdout 13/02/24 13:33:02 WARN mapred.JobClient: Error reading task outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000002_2&filter=stderr 13/02/24 13:33:08 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_m_000001_0, Status : FAILED Error initializing attempt_201301141457_0034_m_000001_0: java.lang.InternalError at sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769) at sun.misc.URLClassPath.getResource(URLClassPath.java:185) at sun.misc.URLClassPath.getResource(URLClassPath.java:237) at java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113) at java.lang.ClassLoader.getResource(ClassLoader.java:974) at java.lang.ClassLoader.getResource(ClassLoader.java:972) at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075) at java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181) at java.lang.Class.getResourceAsStream(Class.java:2045) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370) at java.security.AccessController.doPrivileged(Native Method) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:250) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:241) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607) at org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310) at org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559) at org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:416) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121) at org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1203) at org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1118) at org.apache.hadoop.mapred.TaskTracker$5.run(TaskTracker.java:2430) at java.lang.Thread.run(Thread.java:679) Caused by: java.io.FileNotFoundException: /usr/lib/jvm/java-1.6.0-openjdk-1.6.0.0.x86_64/jre/lib/resources.jar at sun.misc.URLClassPath$JarLoader.getJarFile(URLClassPath.java:658) at sun.misc.URLClassPath$JarLoader.access$600(URLClassPath.java:555) at sun.misc.URLClassPath$JarLoader$1.run(URLClassPath.java:622) at java.security.AccessController.doPrivileged(Native Method) at sun.misc.URLClassPath$JarLoader.ensureOpen(URLClassPath.java:614) at sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:767) ... 27 more 13/02/24 13:33:08 WARN mapred.JobClient: Error reading task outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000001_0&filter=stdout 13/02/24 13:33:08 WARN mapred.JobClient: Error reading task outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_m_000001_0&filter=stderr 13/02/24 13:33:11 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_r_000001_0, Status : FAILED Error initializing attempt_201301141457_0034_r_000001_0: java.lang.InternalError at sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:769) at sun.misc.URLClassPath.getResource(URLClassPath.java:185) at sun.misc.URLClassPath.getResource(URLClassPath.java:237) at java.lang.ClassLoader.getBootstrapResource(ClassLoader.java:1113) at java.lang.ClassLoader.getResource(ClassLoader.java:974) at java.lang.ClassLoader.getResource(ClassLoader.java:972) at java.lang.ClassLoader.getSystemResource(ClassLoader.java:1075) at java.lang.ClassLoader.getSystemResourceAsStream(ClassLoader.java:1181) at java.lang.Class.getResourceAsStream(Class.java:2045) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory$1.run(OutputPropertiesFactory.java:370) at java.security.AccessController.doPrivileged(Native Method) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.loadPropertiesFile(OutputPropertiesFactory.java:366) at com.sun.org.apache.xml.internal.serializer.OutputPropertiesFactory.getDefaultMethodProperties(OutputPropertiesFactory.java:267) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.setDefaults(TransformerImpl.java:1123) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.createOutputProperties(TransformerImpl.java:1084) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:250) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerImpl.<init>(TransformerImpl.java:241) at com.sun.org.apache.xalan.internal.xsltc.trax.TransformerFactoryImpl.newTransformer(TransformerFactoryImpl.java:607) at org.apache.hadoop.conf.Configuration.writeXml(Configuration.java:1310) at org.apache.hadoop.mapred.JobLocalizer.writeLocalJobFile(JobLocalizer.java:559) at org.apache.hadoop.mapred.TaskTracker$4.run(TaskTracker.java:1227) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:416) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121) at org.apache.hadoop.mapred.TaskTracker.initializeJob(TaskTracker.java:1203) at org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:1118) at org.apache.hadoop.mapred.TaskTracker$5.run(TaskTracker.java:2430) at java.lang.Thread.run(Thread.java:679) Caused by: java.io.FileNotFoundException: /usr/lib/jvm/java-1.6.0-openjdk-1.6.0.0.x86_64/jre/lib/resources.jar at sun.misc.URLClassPath$JarLoader.getJarFile(URLClassPath.java:658) at sun.misc.URLClassPath$JarLoader.access$600(URLClassPath.java:555) at sun.misc.URLClassPath$JarLoader$1.run(URLClassPath.java:622) at java.security.AccessController.doPrivileged(Native Method) at sun.misc.URLClassPath$JarLoader.ensureOpen(URLClassPath.java:614) at sun.misc.URLClassPath$JarLoader.getResource(URLClassPath.java:767) ... 27 more 13/02/24 13:33:11 WARN mapred.JobClient: Error reading task outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_r_000001_0&filter=stdout 13/02/24 13:33:11 WARN mapred.JobClient: Error reading task outputhttp://ADUAE045-LAP-V:50060/tasklog?plaintext=true&attemptid=attempt_201301141457_0034_r_000001_0&filter=stderr 13/02/24 13:33:19 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_m_000001_1, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_m_000001_1: execvp: No such file or directory 13/02/24 13:33:25 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_r_000001_1, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_r_000001_1: execvp: No such file or directory 13/02/24 13:33:31 INFO mapred.JobClient: Task Id : attempt_201301141457_0034_m_000001_2, Status : FAILED java.lang.Throwable: Child Error at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271) Caused by: java.io.IOException: Task process exit with nonzero status of 1. at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258) attempt_201301141457_0034_m_000001_2: execvp: No such file or directory 13/02/24 13:33:37 INFO mapred.JobClient: Job complete: job_201301141457_0034 13/02/24 13:33:37 INFO mapred.JobClient: Counters: 4 13/02/24 13:33:37 INFO mapred.JobClient: Job Counters 13/02/24 13:33:37 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=15112 13/02/24 13:33:37 INFO mapred.JobClient: Total time spent by all reduces waiting after reserving slots (ms)=0 13/02/24 13:33:37 INFO mapred.JobClient: Total time spent by all maps waiting after reserving slots (ms)=0 13/02/24 13:33:37 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=9091 On Sat, Feb 23, 2013 at 5:05 PM, Hemanth Yamijala <yhema...@thoughtworks.com > wrote: > Can you try this ? Pick a class like WordCount from your package and > execute this command: > > javap -classpath <path to your jar> -verbose org.myorg.Wordcount | grep > version. > > For e.g. here's what I get for my class: > > $ javap -verbose WCMapper | grep version > minor version: 0 > major version: 50 > > Please paste the output of this - we can verify what the problem is. > > Thanks > Hemanth > > > On Sat, Feb 23, 2013 at 4:45 PM, Fatih Haltas <fatih.hal...@nyu.edu>wrote: > >> Hi again, >> >> Thanks for your help but now, I am struggling with the same problem on a >> machine. As the preivous problem, I just decrease the Java version by Java >> 6, but this time I could not solve the problem. >> >> those are outputs that may explain the situation: >> >> --------------------------------------------------------------------------------------------------------------------------------------------- >> 1. I could not run my own code, to check the system I just tried to run >> basic wordcount example without any modification, except package info. >> ************************************************** >> COMMAND EXECUTED: hadoop jar my.jar org.myorg.WordCount NetFlow >> NetFlow.out >> Warning: $HADOOP_HOME is deprecated. >> >> Exception in thread "main" java.lang.UnsupportedClassVersionError: >> org/myorg/WordCount : Unsupported major.minor version 51.0 >> at java.lang.ClassLoader.defineClass1(Native Method) >> at java.lang.ClassLoader.defineClass(ClassLoader.java:634) >> at >> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >> at java.net.URLClassLoader.defineClass(URLClassLoader.java:277) >> at java.net.URLClassLoader.access$000(URLClassLoader.java:73) >> at java.net.URLClassLoader$1.run(URLClassLoader.java:212) >> at java.security.AccessController.doPrivileged(Native Method) >> at java.net.URLClassLoader.findClass(URLClassLoader.java:205) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:321) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:266) >> at java.lang.Class.forName0(Native Method) >> at java.lang.Class.forName(Class.java:266) >> at org.apache.hadoop.util.RunJar.main(RunJar.java:149) >> >> ************************************************************************************** >> 2. Java version: >> ******************************** >> COMMAND EXECUTED: java -version >> java version "1.6.0_24" >> OpenJDK Runtime Environment (IcedTea6 1.11.6) >> (rhel-1.33.1.11.6.el5_9-x86_64) >> OpenJDK 64-Bit Server VM (build 20.0-b12, mixed mode) >> ********************************** >> 3. JAVA_HOME variable: >> ********************************** >> COMMAND EXECUTED: echo $JAVA_HOME >> /usr/lib/jvm/jre-1.6.0-openjdk.x86_64 >> ******************************************** >> 4. HADOOP version: >> ******************************************* >> COMMAND EXECUTED: hadoop version >> Warning: $HADOOP_HOME is deprecated. >> >> Hadoop 1.0.4 >> Subversion >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r >> 1393290 >> Compiled by hortonfo on Wed Oct 3 05:13:58 UTC 2012 >> From source with checksum fe2baea87c4c81a2c505767f3f9b71f4 >> ******************************************************** >> >> Are these still incompatible with eachother? (Hadoop version and java >> version) >> >> >> Thank you very much. >> >> >> On Tue, Feb 19, 2013 at 10:26 PM, Fatih Haltas <fatih.hal...@nyu.edu>wrote: >> >>> Thank you all very much >>> >>> 19 Şubat 2013 Salı tarihinde Harsh J adlı kullanıcı şöyle yazdı: >>> >>>> Oops. I just noticed Hemanth has been answering on a dupe thread as >>>> well. Lets drop this thread and carry on there :) >>>> >>>> On Tue, Feb 19, 2013 at 11:14 PM, Harsh J <ha...@cloudera.com> wrote: >>>> > Hi, >>>> > >>>> > The new error usually happens if you compile using Java 7 and try to >>>> > run via Java 6 (for example). That is, an incompatibility in the >>>> > runtimes for the binary artifact produced. >>>> > >>>> > On Tue, Feb 19, 2013 at 10:09 PM, Fatih Haltas <fatih.hal...@nyu.edu> >>>> wrote: >>>> >> Thank you very much Harsh, >>>> >> >>>> >> Now, as I promised earlier I am much obliged to you. >>>> >> >>>> >> But, now I solved that problem by just changing the directories then >>>> again >>>> >> creating a jar file of org. but I am getting this error: >>>> >> >>>> >> 1.) What I got >>>> >> >>>> ------------------------------------------------------------------------------ >>>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ hadoop jar >>>> flow19028pm.jar >>>> >> org.myorg.MapReduce /home/hadoop/project/hadoop-data/NetFlow >>>> 19_02.out >>>> >> Warning: $HADOOP_HOME is deprecated. >>>> >> >>>> >> Exception in thread "main" java.lang.UnsupportedClassVersionError: >>>> >> org/myorg/MapReduce : Unsupported major.minor version 51.0 >>>> >> at java.lang.ClassLoader.defineClass1(Native Method) >>>> >> at java.lang.ClassLoader.defineClass(ClassLoader.java:634) >>>> >> at >>>> >> >>>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >>>> >> at >>>> java.net.URLClassLoader.defineClass(URLClassLoader.java:277) >>>> >> at java.net.URLClassLoader.access$000(URLClassLoader.java:73) >>>> >> at java.net.URLClassLoader$1.run(URLClassLoader.java:212) >>>> >> at java.security.AccessController.doPrivileged(Native Method) >>>> >> at java.net.URLClassLoader.findClass(URLClassLoader.java:205) >>>> >> at java.lang.ClassLoader.loadClass(ClassLoader.java:321) >>>> >> at java.lang.ClassLoader.loadClass(ClassLoader.java:266) >>>> >> at java.lang.Class.forName0(Native Method) >>>> >> at java.lang.Class.forName(Class.java:266) >>>> >> at org.apache.hadoop.util.RunJar.main(RunJar.java:149) >>>> >> >>>> >> 2.) How I create my jar >>>> >> >>>> ------------------------------------------------------------------------------------- >>>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar cvf flow19028pm.jar >>>> org >>>> >> added manifest >>>> >> adding: org/(in = 0) (out= 0)(stored 0%) >>>> >> adding: org/myorg/(in = 0) (out= 0)(stored 0%) >>>> >> adding: org/myorg/MapReduce$FlowPortReducer.class(in = 1661) (out= >>>> >> 690)(deflated 58%) >>>> >> adding: org/myorg/MapReduce.class(in = 1587) (out= 903)(deflated 43%) >>>> >> adding: org/myorg/MapReduce$FlowPortMapper.class(in = 1874) (out= >>>> >> 823)(deflated 56%) >>>> >> >>>> >> 3.) Content of my jar file >>>> >> >>>> --------------------------------------------------------------------------------------- >>>> >> [hadoop@ADUAE042-LAP-V flowclasses_18_02]$ jar tf flow19028pm.jar >>>> >> META-INF/ >>>> >> META-INF/MANIFEST.MF >>>> >> org/ >>>> >> org/myorg/ >>>> >> org/myorg/MapReduce$FlowPortReducer.class >>>> >> org/myorg/MapReduce.class >>>> >> org/myorg/MapReduce$FlowPortMapper.class >>>> >> >>>> ----------------------------------------------------------------------------------------- >>>> >> >>>> >> >>>> >> Thank you very much. >>>> >> >>>> >> >>>> >> On Tue, Feb 19, 2013 at 8:20 PM, Harsh J <ha...@cloudera.com> wrote: >>>> >>> >>>> >>> Your point (4) explains the problem. The jar packed structure should >>>> >>> look like the below, and not how it is presently (one extra top >>>> level >>>> >>> dir is present): >>>> >>> >>>> >>> META-INF/ >>>> >>> META-INF/MANIFEST.MF >>>> >>> org/ >>>> >>> org/myorg/ >>>> >>> org/myorg/WordCount.class >>>> >>> org/myorg/WordCount$TokenizerMapper.class >>>> >>> org/myorg/WordCount$IntSumReducer.clas-- >>>> Harsh J >>>> >>> >> >