Hi

It seems both namenodes was active for a period or the standby node process was stopped for long time. Tip: on the standby node try to backup the fsimage and bootstrap that node again. Be careful because if you do a namenode format again you will lost your datas on the hdfs.

"If you have already formatted the NameNode, or are converting a non-HA-enabled cluster to be HA-enabled, you should now copy over the contents of your NameNode metadata directories to the other, unformatted NameNode by running the command "/hdfs namenode -bootstrapStandby/" on the unformatted NameNode. Running this command will also ensure that the JournalNodes (as configured by *dfs.namenode.shared.edits.dir*) contain sufficient edits transactions to be able to start both NameNodes."

Anyway here is a link about other namenode recovery possibilities:
http://blog.cloudera.com/blog/2012/05/namenode-recovery-tools-for-the-hadoop-distributed-file-system/

Andras


On 2014.12.18. 5:11, Sajid Syed wrote:
Hi All,

I have configured CDH4 with HA. It was working fine for some time and now I started seeing this error and namenode had failed over to secondary server.


2014-12-17 08:44:31,847 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join org.apache.hadoop.hdfs.server.namenode.EditLogInputException: Error replaying edit log at offset 0. Expected transaction ID was 1

Reply via email to