Hey,

Moving this to CDH-users (cdh-u...@cloudera.org) list as its CDH4
packaging/deployment specific. (You can subscribe to it via
https://groups.google.com/a/cloudera.org/group/cdh-user). BCC'd
common-user and cc'd you and Marcos too.

MR jobs will see mapred-site.xml to determine what
'cluster'/'framework' they need to use. Hence, you will need to follow
these deployment instructions when using MR2 with YARN as your choice
of MR: 
https://ccp.cloudera.com/display/CDH4DOC/Deploying+MapReduce+v2+%28YARN%29+on+a+Cluster#DeployingMapReducev2%28YARN%29onaCluster-Step1

(Or if you use CM4, just visit YARN service in it and ask it to deploy
configs automatically to all hosts or get you a client config bundle
for self-deployment. Applying those would do this automatically and
remove pains :))

Can you try this and let us know if it works Anil?

On Thu, Jun 14, 2012 at 2:56 AM, anil gupta <anilgupt...@gmail.com> wrote:
> Hi Marcus,
>
> Sorry i forgot to mention that Job history server is installed and running
> and AFAIK resourcemanager is responsible for running MR jobs. Historyserver
> is only used to get info about MR jobs.
>
> Thanks,
> Anil
>
> On Wed, Jun 13, 2012 at 2:04 PM, Marcos Ortiz <mlor...@uci.cu> wrote:
>
>> According to the CDH 4 official documentation, you should install a
>> JobHistory server for your MRv2 (YARN)
>> cluster.
>> https://ccp.cloudera.com/**display/CDH4DOC/Deploying+**
>> MapReduce+v2+%28YARN%29+on+a+**Cluster<https://ccp.cloudera.com/display/CDH4DOC/Deploying+MapReduce+v2+%28YARN%29+on+a+Cluster>
>>
>> How to configure the HistoryServer
>> https://ccp.cloudera.com/**display/CDH4DOC/Deploying+**
>> MapReduce+v2+%28YARN%29+on+a+**Cluster#DeployingMapReducev2%**
>> 28YARN%29onaCluster-Step3<https://ccp.cloudera.com/display/CDH4DOC/Deploying+MapReduce+v2+%28YARN%29+on+a+Cluster#DeployingMapReducev2%28YARN%29onaCluster-Step3>
>>
>>
>>
>>
>> On 06/13/2012 03:16 PM, anil gupta wrote:
>>
>>> Hi All
>>>
>>> I am using cdh4 for running a HBase cluster on CentOs6.0. I have 5
>>> nodes in my cluster(2 Admin Node and 3 DN).
>>> My resourcemanager is up and running and showing that all three DN are
>>> running the nodemanager. HDFS is also working fine and showing 3 DN's.
>>>
>>> But when i fire the pi example job. It starts to run in Local mode.
>>> Here is the console output:
>>> sudo -u hdfs yarn jar /usr/lib/hadoop-mapreduce/**hadoop-mapreduce-
>>> examples.jar pi 10 1000000000
>>> Number of Maps  = 10
>>> Samples per Map = 1000000000
>>> Wrote input for Map #0
>>> Wrote input for Map #1
>>> Wrote input for Map #2
>>> Wrote input for Map #3
>>> Wrote input for Map #4
>>> Wrote input for Map #5
>>> Wrote input for Map #6
>>> Wrote input for Map #7
>>> Wrote input for Map #8
>>> Wrote input for Map #9
>>> Starting Job
>>> 12/06/13 12:03:27 WARN conf.Configuration: session.id is deprecated.
>>> Instead, use dfs.metrics.session-id
>>> 12/06/13 12:03:27 INFO jvm.JvmMetrics: Initializing JVM Metrics with
>>> processName=JobTracker, sessionId=
>>> 12/06/13 12:03:27 INFO util.NativeCodeLoader: Loaded the native-hadoop
>>> library
>>> 12/06/13 12:03:27 WARN mapred.JobClient: Use GenericOptionsParser for
>>> parsing the arguments. Applications should implement Tool for the
>>> same.
>>> 12/06/13 12:03:28 INFO mapred.FileInputFormat: Total input paths to
>>> process : 10
>>> 12/06/13 12:03:29 INFO mapred.JobClient: Running job: job_local_0001
>>> 12/06/13 12:03:29 INFO mapred.LocalJobRunner: OutputCommitter set in
>>> config null
>>> 12/06/13 12:03:29 INFO mapred.LocalJobRunner: OutputCommitter is
>>> org.apache.hadoop.mapred.**FileOutputCommitter
>>> 12/06/13 12:03:29 WARN mapreduce.Counters: Group
>>> org.apache.hadoop.mapred.Task$**Counter is deprecated. Use
>>> org.apache.hadoop.mapreduce.**TaskCounter instead
>>> 12/06/13 12:03:29 INFO util.ProcessTree: setsid exited with exit code
>>> 0
>>> 12/06/13 12:03:29 INFO mapred.Task:  Using ResourceCalculatorPlugin :
>>> org.apache.hadoop.util.**LinuxResourceCalculatorPlugin@**3d46e381
>>> 12/06/13 12:03:29 WARN mapreduce.Counters: Counter name
>>> MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group
>>> name and  BYTES_READ as counter name instead
>>> 12/06/13 12:03:29 INFO mapred.MapTask: numReduceTasks: 1
>>> 12/06/13 12:03:29 INFO mapred.MapTask: io.sort.mb = 100
>>> 12/06/13 12:03:30 INFO mapred.MapTask: data buffer = 79691776/99614720
>>> 12/06/13 12:03:30 INFO mapred.MapTask: record buffer = 262144/327680
>>> 12/06/13 12:03:30 INFO mapred.JobClient:  map 0% reduce 0%
>>> 12/06/13 12:03:35 INFO mapred.LocalJobRunner: Generated 95735000
>>> samples.
>>> 12/06/13 12:03:36 INFO mapred.JobClient:  map 100% reduce 0%
>>> 12/06/13 12:03:38 INFO mapred.LocalJobRunner: Generated 151872000
>>> samples.
>>>
>>> Here is the content of yarn-site.xml:
>>>
>>> <configuration>
>>>   <property>
>>>     <name>yarn.nodemanager.aux-**services</name>
>>>     <value>mapreduce.shuffle</**value>
>>>   </property>
>>>
>>>   <property>
>>>     <name>yarn.nodemanager.aux-**services.mapreduce.shuffle.**
>>> class</name>
>>>     <value>org.apache.hadoop.**mapred.ShuffleHandler</value>
>>>   </property>
>>>
>>>   <property>
>>>     <name>yarn.log-aggregation-**enable</name>
>>>     <value>true</value>
>>>   </property>
>>>
>>>   <property>
>>>     <description>List of directories to store localized files in.</
>>> description>
>>>     <name>yarn.nodemanager.local-**dirs</name>
>>>     <value>/disk/yarn/local</**value>
>>>   </property>
>>>
>>>   <property>
>>>     <description>Where to store container logs.</description>
>>>     <name>yarn.nodemanager.log-**dirs</name>
>>>     <value>/disk/yarn/logs</value>
>>>   </property>
>>>
>>>   <property>
>>>     <description>Where to aggregate logs to.</description>
>>>     <name>yarn.nodemanager.remote-**app-log-dir</name>
>>>     <value>/var/log/hadoop-yarn/**apps</value>
>>>   </property>
>>>
>>>   <property>
>>>     <description>Classpath for typical applications.</description>
>>>      <name>yarn.application.**classpath</name>
>>>      <value>
>>>         $HADOOP_CONF_DIR,
>>>         $HADOOP_COMMON_HOME/*,$HADOOP_**COMMON_HOME/lib/*,
>>>         $HADOOP_HDFS_HOME/*,$HADOOP_**HDFS_HOME/lib/*,
>>>         $HADOOP_MAPRED_HOME/*,$HADOOP_**MAPRED_HOME/lib/*,
>>>         $YARN_HOME/*,$YARN_HOME/lib/*
>>>      </value>
>>>   </property>
>>> <property>
>>>         <name>yarn.resourcemanager.**resource-tracker.address</**name>
>>>         <value>ihub-an-g1:8025</value>
>>> </property>
>>> <property>
>>>         <name>yarn.resourcemanager.**address</name>
>>>         <value>ihub-an-g1:8040</value>
>>> </property>
>>> <property>
>>>         <name>yarn.resourcemanager.**scheduler.address</name>
>>>         <value>ihub-an-g1:8030</value>
>>> </property>
>>> <property>
>>>         <name>yarn.resourcemanager.**admin.address</name>
>>>         <value>ihub-an-g1:8141</value>
>>> </property>
>>> <property>
>>>         <name>yarn.resourcemanager.**webapp.address</name>
>>>         <value>ihub-an-g1:8088</value>
>>> </property>
>>> <property>
>>>         <name>mapreduce.jobhistory.**intermediate-done-dir</name>
>>>         <value>/disk/mapred/**jobhistory/intermediate/done</**value>
>>> </property>
>>> <property>
>>>         <name>>mapreduce.jobhistory.**done-dir</name>
>>>         <value>/disk/mapred/**jobhistory/done</value>
>>> </property>
>>> </configuration>
>>>
>>> Can anyone tell me what is the problem over here? Appreciate your
>>> help.
>>> Thanks,
>>> Anil Gupta
>>>
>>>
>>>
>>>
>> --
>> Marcos Luis Ortíz Valmaseda
>>  Data Engineer&&  Sr. System Administrator at UCI
>>  http://marcosluis2186.**posterous.com<http://marcosluis2186.posterous.com>
>>  http://www.linkedin.com/in/**marcosluis2186<http://www.linkedin.com/in/marcosluis2186>
>>  Twitter: @marcosluis2186
>>
>>
>> 10mo. ANIVERSARIO DE LA CREACION DE LA UNIVERSIDAD DE LAS CIENCIAS
>> INFORMATICAS...
>> CONECTADOS AL FUTURO, CONECTADOS A LA REVOLUCION
>>
>> http://www.uci.cu
>> http://www.facebook.com/**universidad.uci<http://www.facebook.com/universidad.uci>
>> http://www.flickr.com/photos/**universidad_uci<http://www.flickr.com/photos/universidad_uci>
>>
>
>
>
> --
> Thanks & Regards,
> Anil Gupta



-- 
Harsh J

Reply via email to