Hi Abhishek, I didnt mean to ask you whether it returns result or not. I meant that you should check that the classpath is correct. It should have the directories where yarn is installed.
~Anil On Sun, Jul 29, 2012 at 12:23 PM, abhiTowson cal <abhishek.dod...@gmail.com>wrote: > hi anil, > > Hadoop class path is also working fine. > > Regards > Abhishek > > Thanks for > > On Sun, Jul 29, 2012 at 3:20 PM, abhiTowson cal > <abhishek.dod...@gmail.com> wrote: > > Hi Anil, > > Iam using chd4 with yarn. > > > > On Sun, Jul 29, 2012 at 3:17 PM, Anil Gupta <anilgupt...@gmail.com> > wrote: > >> Are you using cdh4? In you cluster are you using yarn or mr1? > >> Check the classpath of Hadoop by Hadoop classpath command. > >> > >> Best Regards, > >> Anil > >> > >> On Jul 29, 2012, at 12:12 PM, abhiTowson cal <abhishek.dod...@gmail.com> > wrote: > >> > >>> HI Anil, > >>> > >>> I have already tried this,but issue could not be resolved. > >>> > >>> Regards > >>> Abhishek > >>> > >>> On Sun, Jul 29, 2012 at 3:05 PM, anil gupta <anilgupt...@gmail.com> > wrote: > >>>> Hi Abhishek, > >>>> > >>>> Once you make sure that whatever Harsh said in the previous email is > >>>> present in the cluster and then also the job runs in Local Mode. Then > try > >>>> running the job with hadoop --config option. > >>>> Refer to this discussion for more detail: > >>>> > https://groups.google.com/a/cloudera.org/forum/#!topic/cdh-user/4tMGfvJFzrg > >>>> > >>>> HTH, > >>>> Anil > >>>> > >>>> On Sun, Jul 29, 2012 at 11:43 AM, Harsh J <ha...@cloudera.com> wrote: > >>>> > >>>>> For a job to get submitted to a cluster, you will need proper client > >>>>> configurations. Have you configured your mapred-site.xml and > >>>>> yarn-site.xml properly inside /etc/hadoop/conf/mapred-site.xml and > >>>>> /etc/hadoop/conf/yarn-site.xml at the client node? > >>>>> > >>>>> On Mon, Jul 30, 2012 at 12:00 AM, abhiTowson cal > >>>>> <abhishek.dod...@gmail.com> wrote: > >>>>>> Hi All, > >>>>>> > >>>>>> I am getting problem that job is running in localrunner rather than > >>>>>> the cluster enviormnent. > >>>>>> And when am running the job i would not be able to see the job id in > >>>>>> the resource manager UI > >>>>>> > >>>>>> Can you please go through the issues and let me know ASAP. > >>>>>> > >>>>>> sudo -u hdfs hadoop jar > >>>>>> /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar teragen > >>>>>> 1000000 /benchmark/teragen/input > >>>>>> 12/07/29 13:35:59 WARN conf.Configuration: session.id is > deprecated. > >>>>>> Instead, use dfs.metrics.session-id > >>>>>> 12/07/29 13:35:59 INFO jvm.JvmMetrics: Initializing JVM Metrics with > >>>>>> processName=JobTracker, sessionId= > >>>>>> 12/07/29 13:35:59 INFO util.NativeCodeLoader: Loaded the > native-hadoop > >>>>> library > >>>>>> 12/07/29 13:35:59 WARN mapred.JobClient: Use GenericOptionsParser > for > >>>>>> parsing the arguments. Applications should implement Tool for the > >>>>>> same. > >>>>>> Generating 1000000 using 1 maps with step of 1000000 > >>>>>> 12/07/29 13:35:59 INFO mapred.JobClient: Running job: job_local_0001 > >>>>>> 12/07/29 13:35:59 INFO mapred.LocalJobRunner: OutputCommitter set in > >>>>> config null > >>>>>> 12/07/29 13:35:59 INFO mapred.LocalJobRunner: OutputCommitter is > >>>>>> org.apache.hadoop.mapred.FileOutputCommitter > >>>>>> 12/07/29 13:35:59 WARN mapreduce.Counters: Group > >>>>>> org.apache.hadoop.mapred.Task$Counter is deprecated. Use > >>>>>> org.apache.hadoop.mapreduce.TaskCounter instead > >>>>>> 12/07/29 13:35:59 INFO util.ProcessTree: setsid exited with exit > code 0 > >>>>>> 12/07/29 13:35:59 INFO mapred.Task: Using ResourceCalculatorPlugin > : > >>>>>> org.apache.hadoop.util.LinuxResourceCalculatorPlugin@47c297a3 > >>>>>> 12/07/29 13:36:00 WARN mapreduce.Counters: Counter name > >>>>>> MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group > >>>>>> name and BYTES_READ as counter name instead > >>>>>> 12/07/29 13:36:00 INFO mapred.MapTask: numReduceTasks: 0 > >>>>>> 12/07/29 13:36:00 INFO mapred.JobClient: map 0% reduce 0% > >>>>>> 12/07/29 13:36:01 INFO mapred.Task: > Task:attempt_local_0001_m_000000_0 > >>>>>> is done. And is in the process of commiting > >>>>>> 12/07/29 13:36:01 INFO mapred.LocalJobRunner: > >>>>>> 12/07/29 13:36:01 INFO mapred.Task: Task > attempt_local_0001_m_000000_0 > >>>>>> is allowed to commit now > >>>>>> 12/07/29 13:36:01 INFO mapred.FileOutputCommitter: Saved output of > >>>>>> task 'attempt_local_0001_m_000000_0' to > >>>>>> hdfs://hadoop-master-1/benchmark/teragen/input > >>>>>> 12/07/29 13:36:01 INFO mapred.LocalJobRunner: > >>>>>> 12/07/29 13:36:01 INFO mapred.Task: Task > 'attempt_local_0001_m_000000_0' > >>>>> done. > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: map 100% reduce 0% > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Job complete: > job_local_0001 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Counters: 19 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: File System Counters > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: FILE: Number of bytes > >>>>> read=142686 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: FILE: Number of bytes > >>>>>> written=220956 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: FILE: Number of read > >>>>> operations=0 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: FILE: Number of large > >>>>>> read operations=0 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: FILE: Number of write > >>>>> operations=0 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: HDFS: Number of bytes > read=0 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: HDFS: Number of bytes > >>>>>> written=100000000 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: HDFS: Number of read > >>>>> operations=1 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: HDFS: Number of large > >>>>>> read operations=0 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: HDFS: Number of write > >>>>> operations=2 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Map-Reduce Framework > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Map input > records=1000000 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Map output > records=1000000 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Input split bytes=82 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Spilled Records=0 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: CPU time spent (ms)=0 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Physical memory (bytes) > >>>>> snapshot=0 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Virtual memory (bytes) > >>>>> snapshot=0 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Total committed heap > >>>>>> usage (bytes)=124715008 > >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: > >>>>>> org.apache.hadoop.mapreduce.lib.input.FileInputFormatCounter > >>>>>> > >>>>>> Regards > >>>>>> Abhishek > >>>>> > >>>>> > >>>>> > >>>>> -- > >>>>> Harsh J > >>>>> > >>>> > >>>> > >>>> > >>>> -- > >>>> Thanks & Regards, > >>>> Anil Gupta > -- Thanks & Regards, Anil Gupta