Hi
It seems both namenodes was active for a period or the standby node
process was stopped for long time.
Tip: on the standby node try to backup the fsimage and bootstrap that
node again. Be careful because if you do a namenode format again you
will lost your datas on the hdfs.
"If you have already formatted the NameNode, or are converting a
non-HA-enabled cluster to be HA-enabled, you should now copy over the
contents of your NameNode metadata directories to the other, unformatted
NameNode by running the command "/hdfs namenode -bootstrapStandby/" on
the unformatted NameNode. Running this command will also ensure that the
JournalNodes (as configured by *dfs.namenode.shared.edits.dir*) contain
sufficient edits transactions to be able to start both NameNodes."
Anyway here is a link about other namenode recovery possibilities:
http://blog.cloudera.com/blog/2012/05/namenode-recovery-tools-for-the-hadoop-distributed-file-system/
Andras
On 2014.12.18. 5:11, Sajid Syed wrote:
Hi All,
I have configured CDH4 with HA. It was working fine for some time and
now I started seeing this error and namenode had failed over to
secondary server.
2014-12-17 08:44:31,847 FATAL
org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode
join
org.apache.hadoop.hdfs.server.namenode.EditLogInputException: Error
replaying edit log at offset 0. Expected transaction ID was 1