Hi, Did you configured your Name Node to store multiple copies of its metadata?.
You can recover your name node in that situation. #hadoop namenode -recover it will ask you whether you want to continue or not, Please follow the instructions. Thanks On Wed, Aug 7, 2013 at 1:44 PM, Manish Bhoge <[email protected]>wrote: > I am not using HA here. All I am trying here is to make a 2 node > cluster. But before that i wanted to make sure that i am setting up > everything right and make the HDFS up on Pseudo distributed mode. However, > I am suspecting a mistake in my /etc/hosts file. As, I have rename the > local host to myhost-1 > > Please suggest. > > ------------------------------ > *From:* Azuryy Yu <[email protected]> > *To:* [email protected]; Manish Bhoge <[email protected]> > *Sent:* Wednesday, 7 August 2013 1:08 PM > *Subject:* Re: Namenode is failing with expception to join > > Manish, > > you stop HDFS then start HDFS on the standby name node right? > > please looked at https://issues.apache.org/jira/browse/HDFS-5058 > > there are two solutions: > 1) start HDFS on the active name node, nor SBN > 2) copy {namenode.name.dir}/* to the SBN > > I advice #1. > > > > > On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge > <[email protected]>wrote: > > I have all configuration fine. But whenever i start namenode it fails > with a below exception. No clue where to fix this? > > "2013-08-07 02:56:22,754 FATAL > org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join" > > 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: > Number of files = 1 > 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: > Number of files under construction = 0 > 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: > Image file of size 115 loaded in 0 seconds. > 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: > Loaded image for txid 0 from > /data/1/dfs/nn/current/fsimage_0000000000000000000 > 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: > Reading > org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d > expecting start txid #1 > 2013-08-07 02:56:22,752 INFO > org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding > stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' > to transaction ID 1 > 2013-08-07 02:56:22,753 INFO > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics > system... > 2013-08-07 02:56:22,754 INFO > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system > stopped. > 2013-08-07 02:56:22,754 INFO > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system > shutdown complete.2013-08-07 02:56:22,754 FATAL > org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join > java.io.IOException: There appears to be a gap in the edit log. We expected > txid 1, but got txid 515247. > > > > >
