Hi,

Did you configured your Name Node to store multiple copies of its metadata?.

You can recover your name node in that situation.

#hadoop namenode -recover

it will ask you whether you want to continue or not, Please follow the
instructions.

Thanks
On Wed, Aug 7, 2013 at 1:44 PM, Manish Bhoge <[email protected]>wrote:

>  I am not using HA here. All I am trying here is to make a 2 node
> cluster. But before that i wanted to make sure that i am setting up
> everything right and make the HDFS up on Pseudo distributed mode. However,
> I am suspecting a mistake in my /etc/hosts file. As, I have rename the
> local host to myhost-1
>
> Please suggest.
>
>   ------------------------------
> *From:* Azuryy Yu <[email protected]>
> *To:* [email protected]; Manish Bhoge <[email protected]>
> *Sent:* Wednesday, 7 August 2013 1:08 PM
> *Subject:* Re: Namenode is failing with expception to join
>
>  Manish,
>
> you stop HDFS then start HDFS on the standby name node right?
>
> please looked at https://issues.apache.org/jira/browse/HDFS-5058
>
> there are two solutions:
> 1) start HDFS on the active name node, nor SBN
> 2) copy {namenode.name.dir}/* to the SBN
>
> I advice #1.
>
>
>
>
> On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge 
> <[email protected]>wrote:
>
>  I have all configuration fine. But whenever i start namenode it fails
> with a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: 
> Number of files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: 
> Number of files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: 
> Image file of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: 
> Loaded image for txid 0 from 
> /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: 
> Reading 
> org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d 
> expecting start txid #1
> 2013-08-07 02:56:22,752 INFO 
> org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding 
> stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' 
> to transaction ID 1
> 2013-08-07 02:56:22,753 INFO 
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics 
> system...
> 2013-08-07 02:56:22,754 INFO 
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system 
> stopped.
> 2013-08-07 02:56:22,754 INFO 
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system 
> shutdown complete.2013-08-07 02:56:22,754 FATAL 
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected 
> txid 1, but got txid 515247.
>
>
>
>
>

Reply via email to