Ah yes, I overlooked this. Then please check the file are there or not: "ls /home/hadoop/project/hadoop-data/dfs/name"?
Regards, *Stanley Shi,* On Tue, Mar 18, 2014 at 2:06 PM, Azuryy Yu <azury...@gmail.com> wrote: > I don't think this is the case, because there is; > <property> > <name>hadoop.tmp.dir</name> > <value>/home/hadoop/project/hadoop-data</value> > </property> > > > On Tue, Mar 18, 2014 at 1:55 PM, Stanley Shi <s...@gopivotal.com> wrote: > >> one possible reason is that you didn't set the namenode working >> directory, by default it's in "/tmp" folder; and the "/tmp" folder might >> get deleted by the OS without any notification. If this is the case, I am >> afraid you have lost all your namenode data. >> >> *<property> >> <name>dfs.name.dir</name> >> <value>${hadoop.tmp.dir}/dfs/name</value> >> <description>Determines where on the local filesystem the DFS name node >> should store the name table(fsimage). If this is a comma-delimited >> list >> of directories then the name table is replicated in all of the >> directories, for redundancy. </description> >> </property>* >> >> >> Regards, >> *Stanley Shi,* >> >> >> >> On Sun, Mar 16, 2014 at 5:29 PM, Mirko Kämpf <mirko.kae...@gmail.com>wrote: >> >>> Hi, >>> >>> what is the location of the namenodes fsimage and editlogs? >>> And how much memory has the NameNode. >>> >>> Did you work with a Secondary NameNode or a Standby NameNode for >>> checkpointing? >>> >>> Where are your HDFS blocks located, are those still safe? >>> >>> With this information at hand, one might be able to fix your setup, but >>> do not format the old namenode before >>> all is working with a fresh one. >>> >>> Grab a copy of the maintainance guide: >>> http://shop.oreilly.com/product/0636920025085.do?sortby=publicationDate >>> which helps solving such type of problems as well. >>> >>> Best wishes >>> Mirko >>> >>> >>> 2014-03-16 9:07 GMT+00:00 Fatih Haltas <fatih.hal...@nyu.edu>: >>> >>> Dear All, >>>> >>>> I have just restarted machines of my hadoop clusters. Now, I am trying >>>> to restart hadoop clusters again, but getting error on namenode restart. I >>>> am afraid of loosing my data as it was properly running for more than 3 >>>> months. Currently, I believe if I do namenode formatting, it will work >>>> again, however, data will be lost. Is there anyway to solve this without >>>> losing the data. >>>> >>>> I will really appreciate any help. >>>> >>>> Thanks. >>>> >>>> >>>> ===================== >>>> Here is the logs; >>>> ==================== >>>> 2014-02-26 16:02:39,698 INFO >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG: >>>> /************************************************************ >>>> STARTUP_MSG: Starting NameNode >>>> STARTUP_MSG: host = ADUAE042-LAP-V/127.0.0.1 >>>> STARTUP_MSG: args = [] >>>> STARTUP_MSG: version = 1.0.4 >>>> STARTUP_MSG: build = >>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r >>>> 1393290; compiled by 'hortonfo' on Wed Oct 3 05:13:58 UTC 2012 >>>> ************************************************************/ >>>> 2014-02-26 16:02:40,005 INFO >>>> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from >>>> hadoop-metrics2.properties >>>> 2014-02-26 16:02:40,019 INFO >>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source >>>> MetricsSystem,sub=Stats registered. >>>> 2014-02-26 16:02:40,021 INFO >>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot >>>> period at 10 second(s). >>>> 2014-02-26 16:02:40,021 INFO >>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system >>>> started >>>> 2014-02-26 16:02:40,169 INFO >>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi >>>> registered. >>>> 2014-02-26 16:02:40,193 INFO >>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm >>>> registered. >>>> 2014-02-26 16:02:40,194 INFO >>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source >>>> NameNode registered. >>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: VM type >>>> = 64-bit >>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: 2% max >>>> memory = 17.77875 MB >>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: capacity >>>> = 2^21 = 2097152 entries >>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: >>>> recommended=2097152, actual=2097152 >>>> 2014-02-26 16:02:40,273 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=hadoop >>>> 2014-02-26 16:02:40,273 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup >>>> 2014-02-26 16:02:40,274 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>>> isPermissionEnabled=true >>>> 2014-02-26 16:02:40,279 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>>> dfs.block.invalidate.limit=100 >>>> 2014-02-26 16:02:40,279 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>>> isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s), >>>> accessTokenLifetime=0 min(s) >>>> 2014-02-26 16:02:40,724 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered >>>> FSNamesystemStateMBean and NameNodeMXBean >>>> 2014-02-26 16:02:40,749 INFO >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names >>>> occuring more than 10 times >>>> 2014-02-26 16:02:40,780 ERROR >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem >>>> initialization failed. >>>> java.io.IOException: NameNode is not formatted. >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:388) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:362) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:276) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:496) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1279) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288) >>>> 2014-02-26 16:02:40,781 ERROR >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException: >>>> NameNode is not formatted. >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:388) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:362) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:276) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:496) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1279) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288) >>>> >>>> 2014-02-26 16:02:40,781 INFO >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: >>>> /************************************************************ >>>> SHUTDOWN_MSG: Shutting down NameNode at ADUAE042-LAP-V/127.0.0.1 >>>> ************************************************************/ >>>> >>>> =========================== >>>> Here is the core-site.xml >>>> =========================== >>>> <?xml version="1.0"?> >>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>>> >>>> <!-- Put site-specific property overrides in this file. --> >>>> >>>> <configuration> >>>> <property> >>>> <name>fs.default.name</name> >>>> <value>-BLANKED</value> >>>> </property> >>>> <property> >>>> <name>hadoop.tmp.dir</name> >>>> <value>/home/hadoop/project/hadoop-data</value> >>>> </property> >>>> </configuration> >>>> >>>> >>>> >>>> >>> >> >