Hi Abhishek,

I didnt mean to ask you whether it returns result or not. I meant that you
should check that the classpath is correct. It should have the directories
where yarn is installed.

~Anil

On Sun, Jul 29, 2012 at 12:23 PM, abhiTowson cal
<abhishek.dod...@gmail.com>wrote:

> hi anil,
>
> Hadoop class path is also working fine.
>
> Regards
> Abhishek
>
> Thanks for
>
> On Sun, Jul 29, 2012 at 3:20 PM, abhiTowson cal
> <abhishek.dod...@gmail.com> wrote:
> > Hi Anil,
> >           Iam using chd4 with yarn.
> >
> > On Sun, Jul 29, 2012 at 3:17 PM, Anil Gupta <anilgupt...@gmail.com>
> wrote:
> >> Are you using cdh4? In you cluster are you using yarn or mr1?
> >> Check the classpath of Hadoop by Hadoop classpath command.
> >>
> >> Best Regards,
> >> Anil
> >>
> >> On Jul 29, 2012, at 12:12 PM, abhiTowson cal <abhishek.dod...@gmail.com>
> wrote:
> >>
> >>> HI Anil,
> >>>
> >>> I have already tried this,but issue could not be resolved.
> >>>
> >>> Regards
> >>> Abhishek
> >>>
> >>> On Sun, Jul 29, 2012 at 3:05 PM, anil gupta <anilgupt...@gmail.com>
> wrote:
> >>>> Hi Abhishek,
> >>>>
> >>>> Once you make sure that whatever Harsh said in the previous email is
> >>>> present in the cluster and then also the job runs in Local Mode. Then
> try
> >>>> running the job with hadoop --config option.
> >>>> Refer to this discussion for more detail:
> >>>>
> https://groups.google.com/a/cloudera.org/forum/#!topic/cdh-user/4tMGfvJFzrg
> >>>>
> >>>> HTH,
> >>>> Anil
> >>>>
> >>>> On Sun, Jul 29, 2012 at 11:43 AM, Harsh J <ha...@cloudera.com> wrote:
> >>>>
> >>>>> For a job to get submitted to a cluster, you will need proper client
> >>>>> configurations. Have you configured your mapred-site.xml and
> >>>>> yarn-site.xml properly inside /etc/hadoop/conf/mapred-site.xml and
> >>>>> /etc/hadoop/conf/yarn-site.xml at the client node?
> >>>>>
> >>>>> On Mon, Jul 30, 2012 at 12:00 AM, abhiTowson cal
> >>>>> <abhishek.dod...@gmail.com> wrote:
> >>>>>> Hi All,
> >>>>>>
> >>>>>> I am getting problem that job is running in localrunner rather than
> >>>>>> the cluster enviormnent.
> >>>>>> And when am running the job i would not be able to see the job id in
> >>>>>> the resource manager UI
> >>>>>>
> >>>>>> Can you please go through the issues and let me know ASAP.
> >>>>>>
> >>>>>> sudo -u hdfs hadoop jar
> >>>>>> /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar teragen
> >>>>>> 1000000 /benchmark/teragen/input
> >>>>>> 12/07/29 13:35:59 WARN conf.Configuration: session.id is
> deprecated.
> >>>>>> Instead, use dfs.metrics.session-id
> >>>>>> 12/07/29 13:35:59 INFO jvm.JvmMetrics: Initializing JVM Metrics with
> >>>>>> processName=JobTracker, sessionId=
> >>>>>> 12/07/29 13:35:59 INFO util.NativeCodeLoader: Loaded the
> native-hadoop
> >>>>> library
> >>>>>> 12/07/29 13:35:59 WARN mapred.JobClient: Use GenericOptionsParser
> for
> >>>>>> parsing the arguments. Applications should implement Tool for the
> >>>>>> same.
> >>>>>> Generating 1000000 using 1 maps with step of 1000000
> >>>>>> 12/07/29 13:35:59 INFO mapred.JobClient: Running job: job_local_0001
> >>>>>> 12/07/29 13:35:59 INFO mapred.LocalJobRunner: OutputCommitter set in
> >>>>> config null
> >>>>>> 12/07/29 13:35:59 INFO mapred.LocalJobRunner: OutputCommitter is
> >>>>>> org.apache.hadoop.mapred.FileOutputCommitter
> >>>>>> 12/07/29 13:35:59 WARN mapreduce.Counters: Group
> >>>>>> org.apache.hadoop.mapred.Task$Counter is deprecated. Use
> >>>>>> org.apache.hadoop.mapreduce.TaskCounter instead
> >>>>>> 12/07/29 13:35:59 INFO util.ProcessTree: setsid exited with exit
> code 0
> >>>>>> 12/07/29 13:35:59 INFO mapred.Task:  Using ResourceCalculatorPlugin
> :
> >>>>>> org.apache.hadoop.util.LinuxResourceCalculatorPlugin@47c297a3
> >>>>>> 12/07/29 13:36:00 WARN mapreduce.Counters: Counter name
> >>>>>> MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group
> >>>>>> name and  BYTES_READ as counter name instead
> >>>>>> 12/07/29 13:36:00 INFO mapred.MapTask: numReduceTasks: 0
> >>>>>> 12/07/29 13:36:00 INFO mapred.JobClient:  map 0% reduce 0%
> >>>>>> 12/07/29 13:36:01 INFO mapred.Task:
> Task:attempt_local_0001_m_000000_0
> >>>>>> is done. And is in the process of commiting
> >>>>>> 12/07/29 13:36:01 INFO mapred.LocalJobRunner:
> >>>>>> 12/07/29 13:36:01 INFO mapred.Task: Task
> attempt_local_0001_m_000000_0
> >>>>>> is allowed to commit now
> >>>>>> 12/07/29 13:36:01 INFO mapred.FileOutputCommitter: Saved output of
> >>>>>> task 'attempt_local_0001_m_000000_0' to
> >>>>>> hdfs://hadoop-master-1/benchmark/teragen/input
> >>>>>> 12/07/29 13:36:01 INFO mapred.LocalJobRunner:
> >>>>>> 12/07/29 13:36:01 INFO mapred.Task: Task
> 'attempt_local_0001_m_000000_0'
> >>>>> done.
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:  map 100% reduce 0%
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Job complete:
> job_local_0001
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient: Counters: 19
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:   File System Counters
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     FILE: Number of bytes
> >>>>> read=142686
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     FILE: Number of bytes
> >>>>>> written=220956
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     FILE: Number of read
> >>>>> operations=0
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     FILE: Number of large
> >>>>>> read operations=0
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     FILE: Number of write
> >>>>> operations=0
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     HDFS: Number of bytes
> read=0
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     HDFS: Number of bytes
> >>>>>> written=100000000
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     HDFS: Number of read
> >>>>> operations=1
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     HDFS: Number of large
> >>>>>> read operations=0
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     HDFS: Number of write
> >>>>> operations=2
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:   Map-Reduce Framework
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     Map input
> records=1000000
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     Map output
> records=1000000
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     Input split bytes=82
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     Spilled Records=0
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     CPU time spent (ms)=0
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     Physical memory (bytes)
> >>>>> snapshot=0
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     Virtual memory (bytes)
> >>>>> snapshot=0
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:     Total committed heap
> >>>>>> usage (bytes)=124715008
> >>>>>> 12/07/29 13:36:02 INFO mapred.JobClient:
> >>>>>> org.apache.hadoop.mapreduce.lib.input.FileInputFormatCounter
> >>>>>>
> >>>>>> Regards
> >>>>>> Abhishek
> >>>>>
> >>>>>
> >>>>>
> >>>>> --
> >>>>> Harsh J
> >>>>>
> >>>>
> >>>>
> >>>>
> >>>> --
> >>>> Thanks & Regards,
> >>>> Anil Gupta
>



-- 
Thanks & Regards,
Anil Gupta
  • Fwd: abhiTowson cal
    • Re: Harsh J
      • Re: anil gupta
        • Re: abhiTowson cal
          • Re: Anil Gupta
            • Re: abhiTowson cal
              • Re: abhiTowson cal
                • Re: anil gupta
                • Re: abhiTowson cal
                • Re: anil gupta
              • Re: anil gupta
    • Re: abhiTowson cal

Reply via email to