Hi Nitin,
 Even after formatting using hdfs namenode -format, i keep seeing namenode
not formatted in the logs when i try to start namenode........
12/24 20:33:26 INFO namenode.FSNamesystem: supergroup=supergroup
13/12/24 20:33:26 INFO namenode.FSNamesystem: isPermissionEnabled=true
13/12/24 20:33:26 INFO namenode.NameNode: Caching file names occuring more
than 10 times
13/12/24 20:33:26 INFO namenode.NNStorage: Storage directory
/usr/local/Software/hadoop-0.23.10/data/hdfs/namenode has been successfully
formatted.
13/12/24 20:33:26 INFO namenode.FSImage: Saving image file
/usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/current/fsimage.ckpt_0000000000000000000
using no compression
13/12/24 20:33:26 INFO namenode.FSImage: Image file of size 124 saved in 0
seconds.
13/12/24 20:33:26 INFO namenode.NNStorageRetentionManager: Going to retain
1 images with txid >= 0
13/12/24 20:33:26 INFO util.ExitUtil: Exiting with status 0
13/12/24 20:33:26 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
************************************************************/


2013-12-24 20:33:46,337 INFO org.apache.hadoop.hdfs.server.common.Storage:
Lock on /usr/local/Software/hadoop-2.2.0/data/hdfs/namenode/in_use.lock
acquired by nodename [email protected]
2013-12-24 20:33:46,339 INFO org.mortbay.log: Stopped
[email protected]:50070
2013-12-24 20:33:46,340 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode
metrics system...
2013-12-24 20:33:46,340 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
stopped.
2013-12-24 20:33:46,340 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
shutdown complete.
2013-12-24 20:33:46,340 FATAL
org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
java.io.IOException: NameNode is not formatted.
        at
org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:210)
        at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787)
        at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568)
        at
org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443)
        at
org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491)
        at
org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684)
        at
org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669)
        at
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254)
        at
org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320)
2013-12-24 20:33:46,342 INFO org.apache.hadoop.util.ExitUtil: Exiting with
status 1
2013-12-24 20:33:46,343 INFO
org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
************************************************************/




On Tue, Dec 24, 2013 at 3:13 PM, Nitin Pawar <[email protected]>wrote:

> the issue here is you tried one version of hadoop and then changed to a
> different version.
>
> You can not do that directly with hadoop. You need to follow a process
> while upgrading hadoop versions.
>
> For now as you are just starting with hadoop, I would recommend just run a
> dfs format and start the hdfs again
>
>
> On Tue, Dec 24, 2013 at 2:57 PM, Sitaraman Vilayannur <
> [email protected]> wrote:
>
>> When i run namenode with upgrade option i get the following error and
>> and namenode dosent start...
>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange:
>> STATE* Network topology has 0 racks and 0 datanodes
>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange:
>> STATE* UnderReplicatedBlocks has 0 blocks
>> 2013-12-24 14:48:38,631 INFO org.apache.hadoop.ipc.Server: IPC Server
>> Responder: starting
>> 2013-12-24 14:48:38,632 INFO org.apache.hadoop.ipc.Server: IPC Server
>> listener on 9000: starting
>> 2013-12-24 14:48:38,633 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: NameNode RPC up at:
>> 192.168.1.2/192.168.1.2:9000
>> 2013-12-24 14:48:38,633 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Starting services
>> required for active state
>> 2013-12-24 14:50:50,060 ERROR
>> org.apache.hadoop.hdfs.server.namenode.NameNode: RECEIVED SIGNAL 15:
>> SIGTERM
>> 2013-12-24 14:50:50,062 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>> /************************************************************
>> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
>> ************************************************************/
>>
>>
>> On 12/24/13, Sitaraman Vilayannur <[email protected]> wrote:
>> > Found it,
>> >  I get the following error on starting namenode in 2.2
>> >
>> 10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar
>> > STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common
>> > -r 1529768; compiled by 'hortonmu' on 2013-10-07T06:28Z
>> > STARTUP_MSG:   java = 1.7.0_45
>> > ************************************************************/
>> > 2013-12-24 13:25:48,876 INFO
>> > org.apache.hadoop.hdfs.server.namenode.NameNode: registered UNIX
>> > signal handlers for [TERM, HUP, INT]
>> > 2013-12-24 13:25:49,042 INFO
>> > org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>> > hadoop-metrics2.properties
>> > 2013-12-24 13:25:49,102 INFO
>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
>> > period at 10 second(s).
>> > 2013-12-24 13:25:49,102 INFO
>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
>> > system started
>> > 2013-12-24 13:25:49,232 WARN org.apache.hadoop.util.NativeCodeLoader:
>> > Unable to load native-hadoop library for your platform... using
>> > builtin-java classes where applicable
>> > 2013-12-24 13:25:49,375 INFO org.mortbay.log: Logging to
>> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>> > org.mortbay.log.Slf4jLog
>> > 2013-12-24 13:25:49,410 INFO org.apache.hadoop.http.HttpServer: Added
>> > global filter 'safety'
>> > (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
>> > filter static_user_filter
>> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
>> > to context hdfs
>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
>> > filter static_user_filter
>> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
>> > to context static
>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
>> > filter static_user_filter
>> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
>> > to context logs
>> > 2013-12-24 13:25:49,422 INFO org.apache.hadoop.http.HttpServer:
>> > dfs.webhdfs.enabled = false
>> > 2013-12-24 13:25:49,432 INFO org.apache.hadoop.http.HttpServer: Jetty
>> > bound to port 50070
>> > 2013-12-24 13:25:49,432 INFO org.mortbay.log: jetty-6.1.26
>> > 2013-12-24 13:25:49,459 WARN org.mortbay.log: Can't reuse
>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08, using
>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08_2787234685293301311
>> > 2013-12-24 13:25:49,610 INFO org.mortbay.log: Started
>> > [email protected]:50070
>> > 2013-12-24 13:25:49,611 INFO
>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
>> > 0.0.0.0:50070
>> > 2013-12-24 13:25:49,628 WARN
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one image
>> > storage directory (dfs.namenode.name.dir) configured. Beware of
>> > dataloss due to lack of redundant storage directories!
>> > 2013-12-24 13:25:49,628 WARN
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one
>> > namespace edits storage directory (dfs.namenode.edits.dir) configured.
>> > Beware of dataloss due to lack of redundant storage directories!
>> > 2013-12-24 13:25:49,668 INFO
>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read includes:
>> > HostSet(
>> > )
>> > 2013-12-24 13:25:49,669 INFO
>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read excludes:
>> > HostSet(
>> > )
>> > 2013-12-24 13:25:49,670 INFO
>> > org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager:
>> > dfs.block.invalidate.limit=1000
>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: Computing
>> > capacity for map BlocksMap
>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: VM type
>> =
>> > 64-bit
>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: 2.0% max
>> > memory = 889 MB
>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: capacity
>> >  = 2^21 = 2097152 entries
>> > 2013-12-24 13:25:49,677 INFO
>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>> > dfs.block.access.token.enable=false
>> > 2013-12-24 13:25:49,677 INFO
>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>> > defaultReplication         = 1
>> > 2013-12-24 13:25:49,677 INFO
>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>> > maxReplication             = 512
>> > 2013-12-24 13:25:49,677 INFO
>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>> > minReplication             = 1
>> > 2013-12-24 13:25:49,677 INFO
>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>> > maxReplicationStreams      = 2
>> > 2013-12-24 13:25:49,677 INFO
>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>> > shouldCheckForEnoughRacks  = false
>> > 2013-12-24 13:25:49,677 INFO
>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>> > replicationRecheckInterval = 3000
>> > 2013-12-24 13:25:49,677 INFO
>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>> > encryptDataTransfer        = false
>> > 2013-12-24 13:25:49,681 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner
>> >   = sitaraman (auth:SIMPLE)
>> > 2013-12-24 13:25:49,681 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup
>> >   = supergroup
>> > 2013-12-24 13:25:49,681 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> > isPermissionEnabled = true
>> > 2013-12-24 13:25:49,681 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: HA Enabled: false
>> > 2013-12-24 13:25:49,682 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Append Enabled:
>> > true
>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: Computing
>> > capacity for map INodeMap
>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: VM type
>> =
>> > 64-bit
>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: 1.0% max
>> > memory = 889 MB
>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: capacity
>> >  = 2^20 = 1048576 entries
>> > 2013-12-24 13:25:49,802 INFO
>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names
>> > occuring more than 10 times
>> > 2013-12-24 13:25:49,804 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> > dfs.namenode.safemode.threshold-pct = 0.9990000128746033
>> > 2013-12-24 13:25:49,804 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> > dfs.namenode.safemode.min.datanodes = 0
>> > 2013-12-24 13:25:49,804 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> > dfs.namenode.safemode.extension     = 30000
>> > 2013-12-24 13:25:49,805 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache on
>> > namenode is enabled
>> > 2013-12-24 13:25:49,805 INFO
>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache will
>> > use 0.03 of total heap and retry cache entry expiry time is 600000
>> > millis
>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: Computing
>> > capacity for map Namenode Retry Cache
>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: VM type
>> =
>> > 64-bit
>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet:
>> > 0.029999999329447746% max memory = 889 MB
>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: capacity
>> >  = 2^15 = 32768 entries
>> > 2013-12-24 13:25:49,816 INFO
>> > org.apache.hadoop.hdfs.server.common.Storage: Lock on
>> > /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/in_use.lock
>> > acquired by nodename [email protected]
>> > 2013-12-24 13:25:49,861 INFO org.mortbay.log: Stopped
>> > [email protected]:50070
>> > 2013-12-24 13:25:49,964 INFO
>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode
>> > metrics system...
>> > 2013-12-24 13:25:49,965 INFO
>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
>> > system stopped.
>> > 2013-12-24 13:25:49,965 INFO
>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
>> > system shutdown complete.
>> > 2013-12-24 13:25:49,965 FATAL
>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode
>> > join
>> > java.io.IOException:
>> > File system image contains an old layout version -39.
>> > An upgrade to version -47 is required.
>> > Please restart NameNode with -upgrade option.
>> >       at
>> >
>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:221)
>> >       at
>> >
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787)
>> >       at
>> >
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568)
>> >       at
>> >
>> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443)
>> >       at
>> >
>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491)
>> >       at
>> >
>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684)
>> >       at
>> >
>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669)
>> >       at
>> >
>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254)
>> >       at
>> > org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320)
>> > 2013-12-24 13:25:49,967 INFO org.apache.hadoop.util.ExitUtil: Exiting
>> > with status 1
>> > 2013-12-24 13:25:49,968 INFO
>> > org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>> > /************************************************************
>> > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
>> > ************************************************************/
>> >
>> > On 12/24/13, Sitaraman Vilayannur <[email protected]>
>> wrote:
>> >> The line beginning with ulimit that i have appended below, i thought
>> >> was the log file?
>> >>
>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
>> >> Sitaraman
>> >> On 12/24/13, Nitin Pawar <[email protected]> wrote:
>> >>> Without log, very hard to guess what's happening.
>> >>>
>> >>> Can you clean up the log directory and then start over and check for
>> the
>> >>> logs again.
>> >>>
>> >>>
>> >>> On Tue, Dec 24, 2013 at 11:44 AM, Sitaraman Vilayannur <
>> >>> [email protected]> wrote:
>> >>>
>> >>>> Hi Nitin,
>> >>>>  I moved to the release 2.2.0 on starting node manager it remains
>> >>>> silent without errors but nodemanager dosent start....while it does
>> in
>> >>>> the earlier 0.23 version
>> >>>>
>> >>>>
>> >>>> ./hadoop-daemon.sh start namenode
>> >>>> starting namenode, logging to
>> >>>>
>> >>>>
>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
>> >>>> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded library
>> >>>> /usr/local/Software/hadoop-2.2.0/lib/native/libhadoop.so.1.0.0 which
>> >>>> might have disabled stack guard. The VM will try to fix the stack
>> >>>> guard now.
>> >>>> It's highly recommended that you fix the library with 'execstack -c
>> >>>> <libfile>', or link it with '-z noexecstack'.
>> >>>> [sitaraman@localhost sbin]$ jps
>> >>>> 13444 Jps
>> >>>> [sitaraman@localhost sbin]$ vi
>> >>>>
>> >>>>
>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
>> >>>>
>> >>>>
>> >>>> ulimit -a for user sitaraman
>> >>>> core file size          (blocks, -c) 0
>> >>>> data seg size           (kbytes, -d) unlimited
>> >>>> scheduling priority             (-e) 0
>> >>>> file size               (blocks, -f) unlimited
>> >>>> pending signals                 (-i) 135104
>> >>>> max locked memory       (kbytes, -l) 32
>> >>>> max memory size         (kbytes, -m) unlimited
>> >>>> open files                      (-n) 1024
>> >>>> pipe size            (512 bytes, -p) 8
>> >>>> POSIX message queues     (bytes, -q) 819200
>> >>>> real-time priority              (-r) 0
>> >>>> stack size              (kbytes, -s) 10240
>> >>>> cpu time               (seconds, -t) unlimited
>> >>>> max user processes              (-u) 135104
>> >>>> virtual memory          (kbytes, -v) unlimited
>> >>>> file locks                      (-x) unlimited
>> >>>>
>> >>>>
>> >>>> On 12/24/13, Nitin Pawar <[email protected]> wrote:
>> >>>> > For now you can ignore this warning,
>> >>>> > it was your first program so you can try building other things and
>> >>>> > slowly
>> >>>> > run the commands mentioned the log message to fix these small
>> >>>> > warnings.
>> >>>> >
>> >>>> >
>> >>>> > On Tue, Dec 24, 2013 at 10:07 AM, Sitaraman Vilayannur <
>> >>>> > [email protected]> wrote:
>> >>>> >
>> >>>> >> Thanks Nitin, That worked,
>> >>>> >> When i run the Pi example, i get the following warning at the end,
>> >>>> >> what must i do about this warning....thanks much for your help.
>> >>>> >> Sitaraman
>> >>>> >> inished in 20.82 seconds
>> >>>> >> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded library
>> >>>> >> /usr/local/Software/hadoop-0.23.10/lib/native/libhadoop.so.1.0.0
>> >>>> >> which
>> >>>> >> might have disabled stack guard. The VM will try to fix the stack
>> >>>> >> guard now.
>> >>>> >> It's highly recommended that you fix the library with 'execstack
>> -c
>> >>>> >> <libfile>', or link it with '-z noexecstack'.
>> >>>> >> 13/12/24 10:05:19 WARN util.NativeCodeLoader: Unable to load
>> >>>> >> native-hadoop library for your platform... using builtin-java
>> >>>> >> classes
>> >>>> >> where applicable
>> >>>> >> Estimated value of Pi is 3.14127500000000000000
>> >>>> >> [sitaraman@localhost mapreduce]$
>> >>>> >>
>> >>>> >> On 12/23/13, Nitin Pawar <[email protected]> wrote:
>> >>>> >> > Can you try starting the process as non root user.
>> >>>> >> > Give proper permissions to the user and start it as a different
>> >>>> >> > user.
>> >>>> >> >
>> >>>> >> > Thanks,
>> >>>> >> > Nitin
>> >>>> >> >
>> >>>> >> >
>> >>>> >> > On Mon, Dec 23, 2013 at 2:15 PM, Sitaraman Vilayannur <
>> >>>> >> > [email protected]> wrote:
>> >>>> >> >
>> >>>> >> >> Hi,
>> >>>> >> >>  When i attempt to start nodemanager i get the following error.
>> >>>> >> >> Any
>> >>>> >> >> help
>> >>>> >> >> appreciated.   I was able to start resource manager datanode,
>> >>>> namenode
>> >>>> >> >> and
>> >>>> >> >> secondarynamenode,
>> >>>> >> >>
>> >>>> >> >>
>> >>>> >> >>    ./yarn-daemon.sh start nodemanager
>> >>>> >> >> starting nodemanager, logging to
>> >>>> >> >>
>> >>>> >>
>> >>>>
>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out
>> >>>> >> >> Unrecognized option: -jvm
>> >>>> >> >> Error: Could not create the Java Virtual Machine.
>> >>>> >> >> Error: A fatal exception has occurred. Program will exit.
>> >>>> >> >> [root@localhost sbin]# emacs
>> >>>> >> >>
>> >>>> >>
>> >>>>
>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out
>> >>>> >> >> &
>> >>>> >> >> [4] 29004
>> >>>> >> >> [root@localhost sbin]# jps
>> >>>> >> >> 28402 SecondaryNameNode
>> >>>> >> >> 30280 Jps
>> >>>> >> >> 28299 DataNode
>> >>>> >> >> 6729 Main
>> >>>> >> >> 26044 ResourceManager
>> >>>> >> >> 28197 NameNode
>> >>>> >> >>
>> >>>> >> >
>> >>>> >> >
>> >>>> >> >
>> >>>> >> > --
>> >>>> >> > Nitin Pawar
>> >>>> >> >
>> >>>> >>
>> >>>> >
>> >>>> >
>> >>>> >
>> >>>> > --
>> >>>> > Nitin Pawar
>> >>>> >
>> >>>>
>> >>>
>> >>>
>> >>>
>> >>> --
>> >>> Nitin Pawar
>> >>>
>> >>
>> >
>>
>
>
>
> --
> Nitin Pawar
>

Reply via email to