Ah yes, I overlooked this. Then please check the file are there or not: "ls
/home/hadoop/project/hadoop-data/dfs/name"?

Regards,
*Stanley Shi,*



On Tue, Mar 18, 2014 at 2:06 PM, Azuryy Yu <azury...@gmail.com> wrote:

> I don't think this is the case, because there is;
>   <property>
>     <name>hadoop.tmp.dir</name>
>     <value>/home/hadoop/project/hadoop-data</value>
>   </property>
>
>
> On Tue, Mar 18, 2014 at 1:55 PM, Stanley Shi <s...@gopivotal.com> wrote:
>
>> one possible reason is that you didn't set the namenode working
>> directory, by default it's in "/tmp" folder; and the "/tmp" folder might
>> get deleted by the OS without any notification. If this is the case, I am
>> afraid you have lost all your namenode data.
>>
>> *<property>
>>   <name>dfs.name.dir</name>
>>   <value>${hadoop.tmp.dir}/dfs/name</value>
>>   <description>Determines where on the local filesystem the DFS name node
>>       should store the name table(fsimage).  If this is a comma-delimited 
>> list
>>       of directories then the name table is replicated in all of the
>>       directories, for redundancy. </description>
>> </property>*
>>
>>
>> Regards,
>> *Stanley Shi,*
>>
>>
>>
>> On Sun, Mar 16, 2014 at 5:29 PM, Mirko Kämpf <mirko.kae...@gmail.com>wrote:
>>
>>> Hi,
>>>
>>> what is the location of the namenodes fsimage and editlogs?
>>> And how much memory has the NameNode.
>>>
>>> Did you work with a Secondary NameNode or a Standby NameNode for
>>> checkpointing?
>>>
>>> Where are your HDFS blocks located, are those still safe?
>>>
>>> With this information at hand, one might be able to fix your setup, but
>>> do not format the old namenode before
>>> all is working with a fresh one.
>>>
>>> Grab a copy of the maintainance guide:
>>> http://shop.oreilly.com/product/0636920025085.do?sortby=publicationDate
>>> which helps solving such type of problems as well.
>>>
>>> Best wishes
>>> Mirko
>>>
>>>
>>> 2014-03-16 9:07 GMT+00:00 Fatih Haltas <fatih.hal...@nyu.edu>:
>>>
>>> Dear All,
>>>>
>>>> I have just restarted machines of my hadoop clusters. Now, I am trying
>>>> to restart hadoop clusters again, but getting error on namenode restart. I
>>>> am afraid of loosing my data as it was properly running for more than 3
>>>> months. Currently, I believe if I do namenode formatting, it will work
>>>> again, however, data will be lost. Is there anyway to solve this without
>>>> losing the data.
>>>>
>>>> I will really appreciate any help.
>>>>
>>>> Thanks.
>>>>
>>>>
>>>> =====================
>>>> Here is the logs;
>>>> ====================
>>>> 2014-02-26 16:02:39,698 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
>>>> /************************************************************
>>>> STARTUP_MSG: Starting NameNode
>>>> STARTUP_MSG:   host = ADUAE042-LAP-V/127.0.0.1
>>>> STARTUP_MSG:   args = []
>>>> STARTUP_MSG:   version = 1.0.4
>>>> STARTUP_MSG:   build =
>>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
>>>> 1393290; compiled by 'hortonfo' on Wed Oct  3 05:13:58 UTC 2012
>>>> ************************************************************/
>>>> 2014-02-26 16:02:40,005 INFO
>>>> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>>>> hadoop-metrics2.properties
>>>> 2014-02-26 16:02:40,019 INFO
>>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>>> MetricsSystem,sub=Stats registered.
>>>> 2014-02-26 16:02:40,021 INFO
>>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
>>>> period at 10 second(s).
>>>> 2014-02-26 16:02:40,021 INFO
>>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
>>>> started
>>>> 2014-02-26 16:02:40,169 INFO
>>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
>>>> registered.
>>>> 2014-02-26 16:02:40,193 INFO
>>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm
>>>> registered.
>>>> 2014-02-26 16:02:40,194 INFO
>>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>>> NameNode registered.
>>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: VM type
>>>>       = 64-bit
>>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: 2% max
>>>> memory = 17.77875 MB
>>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: capacity
>>>>      = 2^21 = 2097152 entries
>>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet:
>>>> recommended=2097152, actual=2097152
>>>> 2014-02-26 16:02:40,273 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=hadoop
>>>> 2014-02-26 16:02:40,273 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
>>>> 2014-02-26 16:02:40,274 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>> isPermissionEnabled=true
>>>> 2014-02-26 16:02:40,279 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>> dfs.block.invalidate.limit=100
>>>> 2014-02-26 16:02:40,279 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>> isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s),
>>>> accessTokenLifetime=0 min(s)
>>>> 2014-02-26 16:02:40,724 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
>>>> FSNamesystemStateMBean and NameNodeMXBean
>>>> 2014-02-26 16:02:40,749 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names
>>>> occuring more than 10 times
>>>> 2014-02-26 16:02:40,780 ERROR
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem
>>>> initialization failed.
>>>> java.io.IOException: NameNode is not formatted.
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:388)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:362)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:276)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:496)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1279)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288)
>>>> 2014-02-26 16:02:40,781 ERROR
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
>>>> NameNode is not formatted.
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:388)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:362)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:276)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:496)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1279)
>>>>         at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288)
>>>>
>>>> 2014-02-26 16:02:40,781 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>>> /************************************************************
>>>> SHUTDOWN_MSG: Shutting down NameNode at ADUAE042-LAP-V/127.0.0.1
>>>> ************************************************************/
>>>>
>>>> ===========================
>>>> Here is the core-site.xml
>>>> ===========================
>>>>  <?xml version="1.0"?>
>>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>>>>
>>>> <!-- Put site-specific property overrides in this file. -->
>>>>
>>>> <configuration>
>>>> <property>
>>>>     <name>fs.default.name</name>
>>>>     <value>-BLANKED</value>
>>>>   </property>
>>>>   <property>
>>>>     <name>hadoop.tmp.dir</name>
>>>>     <value>/home/hadoop/project/hadoop-data</value>
>>>>   </property>
>>>> </configuration>
>>>>
>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to