the issue here is you tried one version of hadoop and then changed to a
different version.

You can not do that directly with hadoop. You need to follow a process
while upgrading hadoop versions.

For now as you are just starting with hadoop, I would recommend just run a
dfs format and start the hdfs again


On Tue, Dec 24, 2013 at 2:57 PM, Sitaraman Vilayannur <
[email protected]> wrote:

> When i run namenode with upgrade option i get the following error and
> and namenode dosent start...
> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange:
> STATE* Network topology has 0 racks and 0 datanodes
> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange:
> STATE* UnderReplicatedBlocks has 0 blocks
> 2013-12-24 14:48:38,631 INFO org.apache.hadoop.ipc.Server: IPC Server
> Responder: starting
> 2013-12-24 14:48:38,632 INFO org.apache.hadoop.ipc.Server: IPC Server
> listener on 9000: starting
> 2013-12-24 14:48:38,633 INFO
> org.apache.hadoop.hdfs.server.namenode.NameNode: NameNode RPC up at:
> 192.168.1.2/192.168.1.2:9000
> 2013-12-24 14:48:38,633 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Starting services
> required for active state
> 2013-12-24 14:50:50,060 ERROR
> org.apache.hadoop.hdfs.server.namenode.NameNode: RECEIVED SIGNAL 15:
> SIGTERM
> 2013-12-24 14:50:50,062 INFO
> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
> ************************************************************/
>
>
> On 12/24/13, Sitaraman Vilayannur <[email protected]> wrote:
> > Found it,
> >  I get the following error on starting namenode in 2.2
> >
> 10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar
> > STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common
> > -r 1529768; compiled by 'hortonmu' on 2013-10-07T06:28Z
> > STARTUP_MSG:   java = 1.7.0_45
> > ************************************************************/
> > 2013-12-24 13:25:48,876 INFO
> > org.apache.hadoop.hdfs.server.namenode.NameNode: registered UNIX
> > signal handlers for [TERM, HUP, INT]
> > 2013-12-24 13:25:49,042 INFO
> > org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
> > hadoop-metrics2.properties
> > 2013-12-24 13:25:49,102 INFO
> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
> > period at 10 second(s).
> > 2013-12-24 13:25:49,102 INFO
> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
> > system started
> > 2013-12-24 13:25:49,232 WARN org.apache.hadoop.util.NativeCodeLoader:
> > Unable to load native-hadoop library for your platform... using
> > builtin-java classes where applicable
> > 2013-12-24 13:25:49,375 INFO org.mortbay.log: Logging to
> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
> > org.mortbay.log.Slf4jLog
> > 2013-12-24 13:25:49,410 INFO org.apache.hadoop.http.HttpServer: Added
> > global filter 'safety'
> > (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
> > filter static_user_filter
> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
> > to context hdfs
> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
> > filter static_user_filter
> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
> > to context static
> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
> > filter static_user_filter
> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
> > to context logs
> > 2013-12-24 13:25:49,422 INFO org.apache.hadoop.http.HttpServer:
> > dfs.webhdfs.enabled = false
> > 2013-12-24 13:25:49,432 INFO org.apache.hadoop.http.HttpServer: Jetty
> > bound to port 50070
> > 2013-12-24 13:25:49,432 INFO org.mortbay.log: jetty-6.1.26
> > 2013-12-24 13:25:49,459 WARN org.mortbay.log: Can't reuse
> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08, using
> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08_2787234685293301311
> > 2013-12-24 13:25:49,610 INFO org.mortbay.log: Started
> > [email protected]:50070
> > 2013-12-24 13:25:49,611 INFO
> > org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
> > 0.0.0.0:50070
> > 2013-12-24 13:25:49,628 WARN
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one image
> > storage directory (dfs.namenode.name.dir) configured. Beware of
> > dataloss due to lack of redundant storage directories!
> > 2013-12-24 13:25:49,628 WARN
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one
> > namespace edits storage directory (dfs.namenode.edits.dir) configured.
> > Beware of dataloss due to lack of redundant storage directories!
> > 2013-12-24 13:25:49,668 INFO
> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read includes:
> > HostSet(
> > )
> > 2013-12-24 13:25:49,669 INFO
> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read excludes:
> > HostSet(
> > )
> > 2013-12-24 13:25:49,670 INFO
> > org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager:
> > dfs.block.invalidate.limit=1000
> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: Computing
> > capacity for map BlocksMap
> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: VM type       =
> > 64-bit
> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: 2.0% max
> > memory = 889 MB
> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: capacity
> >  = 2^21 = 2097152 entries
> > 2013-12-24 13:25:49,677 INFO
> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
> > dfs.block.access.token.enable=false
> > 2013-12-24 13:25:49,677 INFO
> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
> > defaultReplication         = 1
> > 2013-12-24 13:25:49,677 INFO
> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
> > maxReplication             = 512
> > 2013-12-24 13:25:49,677 INFO
> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
> > minReplication             = 1
> > 2013-12-24 13:25:49,677 INFO
> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
> > maxReplicationStreams      = 2
> > 2013-12-24 13:25:49,677 INFO
> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
> > shouldCheckForEnoughRacks  = false
> > 2013-12-24 13:25:49,677 INFO
> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
> > replicationRecheckInterval = 3000
> > 2013-12-24 13:25:49,677 INFO
> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
> > encryptDataTransfer        = false
> > 2013-12-24 13:25:49,681 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner
> >   = sitaraman (auth:SIMPLE)
> > 2013-12-24 13:25:49,681 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup
> >   = supergroup
> > 2013-12-24 13:25:49,681 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
> > isPermissionEnabled = true
> > 2013-12-24 13:25:49,681 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: HA Enabled: false
> > 2013-12-24 13:25:49,682 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Append Enabled:
> > true
> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: Computing
> > capacity for map INodeMap
> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: VM type       =
> > 64-bit
> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: 1.0% max
> > memory = 889 MB
> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: capacity
> >  = 2^20 = 1048576 entries
> > 2013-12-24 13:25:49,802 INFO
> > org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names
> > occuring more than 10 times
> > 2013-12-24 13:25:49,804 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
> > dfs.namenode.safemode.threshold-pct = 0.9990000128746033
> > 2013-12-24 13:25:49,804 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
> > dfs.namenode.safemode.min.datanodes = 0
> > 2013-12-24 13:25:49,804 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
> > dfs.namenode.safemode.extension     = 30000
> > 2013-12-24 13:25:49,805 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache on
> > namenode is enabled
> > 2013-12-24 13:25:49,805 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache will
> > use 0.03 of total heap and retry cache entry expiry time is 600000
> > millis
> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: Computing
> > capacity for map Namenode Retry Cache
> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: VM type       =
> > 64-bit
> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet:
> > 0.029999999329447746% max memory = 889 MB
> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: capacity
> >  = 2^15 = 32768 entries
> > 2013-12-24 13:25:49,816 INFO
> > org.apache.hadoop.hdfs.server.common.Storage: Lock on
> > /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/in_use.lock
> > acquired by nodename [email protected]
> > 2013-12-24 13:25:49,861 INFO org.mortbay.log: Stopped
> > [email protected]:50070
> > 2013-12-24 13:25:49,964 INFO
> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode
> > metrics system...
> > 2013-12-24 13:25:49,965 INFO
> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
> > system stopped.
> > 2013-12-24 13:25:49,965 INFO
> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
> > system shutdown complete.
> > 2013-12-24 13:25:49,965 FATAL
> > org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode
> > join
> > java.io.IOException:
> > File system image contains an old layout version -39.
> > An upgrade to version -47 is required.
> > Please restart NameNode with -upgrade option.
> >       at
> >
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:221)
> >       at
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787)
> >       at
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568)
> >       at
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443)
> >       at
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491)
> >       at
> > org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684)
> >       at
> > org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669)
> >       at
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254)
> >       at
> > org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320)
> > 2013-12-24 13:25:49,967 INFO org.apache.hadoop.util.ExitUtil: Exiting
> > with status 1
> > 2013-12-24 13:25:49,968 INFO
> > org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
> > /************************************************************
> > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
> > ************************************************************/
> >
> > On 12/24/13, Sitaraman Vilayannur <[email protected]>
> wrote:
> >> The line beginning with ulimit that i have appended below, i thought
> >> was the log file?
> >>
> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
> >> Sitaraman
> >> On 12/24/13, Nitin Pawar <[email protected]> wrote:
> >>> Without log, very hard to guess what's happening.
> >>>
> >>> Can you clean up the log directory and then start over and check for
> the
> >>> logs again.
> >>>
> >>>
> >>> On Tue, Dec 24, 2013 at 11:44 AM, Sitaraman Vilayannur <
> >>> [email protected]> wrote:
> >>>
> >>>> Hi Nitin,
> >>>>  I moved to the release 2.2.0 on starting node manager it remains
> >>>> silent without errors but nodemanager dosent start....while it does in
> >>>> the earlier 0.23 version
> >>>>
> >>>>
> >>>> ./hadoop-daemon.sh start namenode
> >>>> starting namenode, logging to
> >>>>
> >>>>
> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
> >>>> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded library
> >>>> /usr/local/Software/hadoop-2.2.0/lib/native/libhadoop.so.1.0.0 which
> >>>> might have disabled stack guard. The VM will try to fix the stack
> >>>> guard now.
> >>>> It's highly recommended that you fix the library with 'execstack -c
> >>>> <libfile>', or link it with '-z noexecstack'.
> >>>> [sitaraman@localhost sbin]$ jps
> >>>> 13444 Jps
> >>>> [sitaraman@localhost sbin]$ vi
> >>>>
> >>>>
> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
> >>>>
> >>>>
> >>>> ulimit -a for user sitaraman
> >>>> core file size          (blocks, -c) 0
> >>>> data seg size           (kbytes, -d) unlimited
> >>>> scheduling priority             (-e) 0
> >>>> file size               (blocks, -f) unlimited
> >>>> pending signals                 (-i) 135104
> >>>> max locked memory       (kbytes, -l) 32
> >>>> max memory size         (kbytes, -m) unlimited
> >>>> open files                      (-n) 1024
> >>>> pipe size            (512 bytes, -p) 8
> >>>> POSIX message queues     (bytes, -q) 819200
> >>>> real-time priority              (-r) 0
> >>>> stack size              (kbytes, -s) 10240
> >>>> cpu time               (seconds, -t) unlimited
> >>>> max user processes              (-u) 135104
> >>>> virtual memory          (kbytes, -v) unlimited
> >>>> file locks                      (-x) unlimited
> >>>>
> >>>>
> >>>> On 12/24/13, Nitin Pawar <[email protected]> wrote:
> >>>> > For now you can ignore this warning,
> >>>> > it was your first program so you can try building other things and
> >>>> > slowly
> >>>> > run the commands mentioned the log message to fix these small
> >>>> > warnings.
> >>>> >
> >>>> >
> >>>> > On Tue, Dec 24, 2013 at 10:07 AM, Sitaraman Vilayannur <
> >>>> > [email protected]> wrote:
> >>>> >
> >>>> >> Thanks Nitin, That worked,
> >>>> >> When i run the Pi example, i get the following warning at the end,
> >>>> >> what must i do about this warning....thanks much for your help.
> >>>> >> Sitaraman
> >>>> >> inished in 20.82 seconds
> >>>> >> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded library
> >>>> >> /usr/local/Software/hadoop-0.23.10/lib/native/libhadoop.so.1.0.0
> >>>> >> which
> >>>> >> might have disabled stack guard. The VM will try to fix the stack
> >>>> >> guard now.
> >>>> >> It's highly recommended that you fix the library with 'execstack -c
> >>>> >> <libfile>', or link it with '-z noexecstack'.
> >>>> >> 13/12/24 10:05:19 WARN util.NativeCodeLoader: Unable to load
> >>>> >> native-hadoop library for your platform... using builtin-java
> >>>> >> classes
> >>>> >> where applicable
> >>>> >> Estimated value of Pi is 3.14127500000000000000
> >>>> >> [sitaraman@localhost mapreduce]$
> >>>> >>
> >>>> >> On 12/23/13, Nitin Pawar <[email protected]> wrote:
> >>>> >> > Can you try starting the process as non root user.
> >>>> >> > Give proper permissions to the user and start it as a different
> >>>> >> > user.
> >>>> >> >
> >>>> >> > Thanks,
> >>>> >> > Nitin
> >>>> >> >
> >>>> >> >
> >>>> >> > On Mon, Dec 23, 2013 at 2:15 PM, Sitaraman Vilayannur <
> >>>> >> > [email protected]> wrote:
> >>>> >> >
> >>>> >> >> Hi,
> >>>> >> >>  When i attempt to start nodemanager i get the following error.
> >>>> >> >> Any
> >>>> >> >> help
> >>>> >> >> appreciated.   I was able to start resource manager datanode,
> >>>> namenode
> >>>> >> >> and
> >>>> >> >> secondarynamenode,
> >>>> >> >>
> >>>> >> >>
> >>>> >> >>    ./yarn-daemon.sh start nodemanager
> >>>> >> >> starting nodemanager, logging to
> >>>> >> >>
> >>>> >>
> >>>>
> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out
> >>>> >> >> Unrecognized option: -jvm
> >>>> >> >> Error: Could not create the Java Virtual Machine.
> >>>> >> >> Error: A fatal exception has occurred. Program will exit.
> >>>> >> >> [root@localhost sbin]# emacs
> >>>> >> >>
> >>>> >>
> >>>>
> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out
> >>>> >> >> &
> >>>> >> >> [4] 29004
> >>>> >> >> [root@localhost sbin]# jps
> >>>> >> >> 28402 SecondaryNameNode
> >>>> >> >> 30280 Jps
> >>>> >> >> 28299 DataNode
> >>>> >> >> 6729 Main
> >>>> >> >> 26044 ResourceManager
> >>>> >> >> 28197 NameNode
> >>>> >> >>
> >>>> >> >
> >>>> >> >
> >>>> >> >
> >>>> >> > --
> >>>> >> > Nitin Pawar
> >>>> >> >
> >>>> >>
> >>>> >
> >>>> >
> >>>> >
> >>>> > --
> >>>> > Nitin Pawar
> >>>> >
> >>>>
> >>>
> >>>
> >>>
> >>> --
> >>> Nitin Pawar
> >>>
> >>
> >
>



-- 
Nitin Pawar

Reply via email to