Hi Andre,

Yups the problem got solved.
The problem I was facing was that JobClient code of my application was messing 
the Hadoop Property:yarn.application.classpath.
After setting it to proper value now things work nice.
Current configuration looks something like this:
yarn.application.classpath=$HADOOP_CONF_DIR, 
$HADOOP_COMMON_HOME/share/hadoop/common/*, 
$HADOOP_COMMON_HOME/share/hadoop/common/lib/*, 
$HADOOP_HDFS_HOME/share/hadoop/hdfs/*, 
$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*, $YARN_HOME/share/hadoop/mapreduce/*, 
$YARN_HOME/share/hadoop/mapreduce/lib/*

Hope this works for you as well….

Cheers,
Subroto Sanyal

On Jul 11, 2012, at 3:14 PM, Andreas Reiter wrote:

> Hi Subroto,
> 
> i have the same problem, can not get my mapreduce jobs to run...
> The container log sais, that org.apache.hadoop.mapreduce.v2.app.MRAppMaster  
> can not be found... :-(
> 
> did you solve it already?
> 
> best regards
> andre
> 
> 
> 
>> ----- Original Message -----
>> From: Subroto <ssan...@datameer.com>
>> Sent: Tue, 5 Jun 2012 14:00:25 +0200
>> To: mapreduce-user@hadoop.apache.org
>> CC:
>> Subject: Re: java.lang.NoClassDefFoundError: 
>> org/apache/hadoop/mapreduce/v2/app/MRAppMaste
> 
>> Hi,
>> 
>> Is it expected to set the /yarn application.classpath/ to:
>> /usr/local/hadoop/etc/hadoop,/usr/local/hadoop/share/hadoop/mapreduce/*,/usr/local/hadoop/share/hadoop/mapreduce/lib/*,/usr/local/hadoop/share/hadoop/common/*,/usr/local/hadoop/share/hadoop/common/lib/*,/usr/local/hadoop/share/hadoop/hdfs/*,/usr/local/hadoop/share/hadoop/hdfs/lib*
>> 
>> I am trying to run the application not from the cluster. Are there any 
>> specific settings needs to be done in Cluster so that I can go ahead with 
>> default /yarn application.classpath?/
>> 
>> Regards,
>> Subroto Sanyal
>> On Jun 5, 2012, at 12:25 PM, Subroto wrote:
>> 
>>> Hi Deva,
>>> 
>>> Tried the yarn application path with absolute values. Still it didn't work. 
>>>  It failed with same stack trace:-(
>>> Now the value of yarn.application.classpath was:
>>> /usr/local/hadoop/etc/hadoop,/usr/local/hadoop/*,/usr/local/hadoop/lib/*,/usr/local/hadoop/*,/usr/local/hadoop/lib/*,/usr/local/hadoop/*,/usr/local/hadoop/lib/*,/usr/local/hadoop/*,/usr/local/hadoop/*
>>> 
>>> Cheers,
>>> Subroto Sanyal
>>> On Jun 5, 2012, at 12:07 PM, Devaraj k wrote:
>>> 
>>>> Hi Subroto,
>>>> 
>>>>    It will not use yarn-env.sh for launching the application master. NM 
>>>> uses the environment set by the client for launching application master.  
>>>> Can you set the environment variables in /etc/profile or update the yarn 
>>>> application classpath with the absolute paths.
>>>> 
>>>> Thanks
>>>> Devaraj
>>>> ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
>>>> *From:*Subroto [ssan...@datameer.com <mailto:ssan...@datameer.com>]
>>>> *Sent:*Tuesday, June 05, 2012 2:25 PM
>>>> *To:*mapreduce-user@hadoop.apache.org 
>>>> <mailto:mapreduce-user@hadoop.apache.org>
>>>> *Subject:*Re: java.lang.NoClassDefFoundError: 
>>>> org/apache/hadoop/mapreduce/v2/app/MRAppMaste
>>>> 
>>>> Hi Deva,
>>>> 
>>>> Thanks for your response.
>>>> The file etc/hadoop/yarn-env.sh has the following entries:
>>>> export HADOOP_MAPRED_HOME=/usr/local/hadoop
>>>> export HADOOP_COMMON_HOME=/usr/local/hadoop
>>>> export HADOOP_HDFS_HOME=/usr/local/hadoop
>>>> export YARN_HOME=/usr/local/hadoop
>>>> export HADOOP_CONF_DIR=/usr/local/hadoop/etc/hadoop
>>>> export YARN_CONF_DIR=$HADOOP_CONF_DIR
>>>> 
>>>> 
>>>> Is it expected to have these variables in profile file of the Linux user??
>>>> 
>>>> I am not using Windows client. My client is running on Mac and the cluster 
>>>> is running on Linux versions.
>>>> 
>>>> Cheers,
>>>> Subroto Sanyal
>>>> On Jun 5, 2012, at 10:50 AM, Devaraj k wrote:
>>>> 
>>>>> Can you check all the hadoop environment variables are set properly in 
>>>>> which the app master is getting launching.
>>>>> 
>>>>> If you are submitting from windows, this might be the 
>>>>> issuehttps://issues.apache.org/jira/browse/MAPREDUCE-4052.
>>>>> 
>>>>> Thanks
>>>>> Devaraj
>>>>> ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
>>>>> *From:*Subroto [ssan...@datameer.com <mailto:ssan...@datameer.com>]
>>>>> *Sent:*Tuesday, June 05, 2012 2:14 PM
>>>>> *To:*mapreduce-user@hadoop.apache.org 
>>>>> <mailto:mapreduce-user@hadoop.apache.org>
>>>>> *Subject:*java.lang.NoClassDefFoundError: 
>>>>> org/apache/hadoop/mapreduce/v2/app/MRAppMaste
>>>>> 
>>>>> Hi,
>>>>> 
>>>>> While running MR Jobs over a yarn cluster I  keep on getting:
>>>>> Exception in thread "main" java.lang.NoClassDefFoundError: 
>>>>> org/apache/hadoop/mapreduce/v2/app/MRAppMaster
>>>>> Caused by: java.lang.ClassNotFoundException: 
>>>>> org.apache.hadoop.mapreduce.v2.app.MRAppMaster
>>>>>   at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
>>>>>   at java.security.AccessController.doPrivileged(Native Method)
>>>>>   at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
>>>>>   at java.lang.ClassLoader.loadClass(ClassLoader.java:307)
>>>>>   at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
>>>>>   at java.lang.ClassLoader.loadClass(ClassLoader.java:248)
>>>>> Could not find the main class: 
>>>>> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
>>>>> 
>>>>> My client is running from a different environment from where the cluster 
>>>>> is running.
>>>>> If I submit a job from the cluster environment; it runs successfully.
>>>>> 
>>>>> I have verified the property yarn.application.classpathbefore submitting 
>>>>> it from the client. The value is set to:
>>>>> $HADOOP_CONF_DIR,$HADOOP_COMMON_HOME/*,$HADOOP_COMMON_HOME/lib/*,$HADOOP_HDFS_HOME/*,$HADOOP_HDFS_HOME/lib/*,$HADOOP_MAPRED_HOME/*,$HADOOP_MAPRED_HOME/lib/*,$YARN_HOME/*,$YARN_HOME/lib/*
>>>>> 
>>>>> Please let me know if I am missing anything.
>>>>> 
>>>>> Cheers,
>>>>> Subroto Sanyal
>>> 
>> 
> 
> 

Reply via email to