[
https://issues.apache.org/jira/browse/HDFS-3071?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13233193#comment-13233193
]
Todd Lipcon commented on HDFS-3071:
-----------------------------------
I tested this manually in addition to the unit tests. For the manual test, I
put one of the NNs in safemode and then issued the failover command:
{code}
todd@todd-w510:~/git/hadoop-common/hadoop-dist/target/hadoop-0.24.0-SNAPSHOT$
./bin/hdfs haadmin -failover nn2 nn1
Failover failed: todd-w510/127.0.0.1:8021 is not ready to become active: Not
ready to go active, since the node is in safemode. Use "hdfs dfsadmin -safemode
leave" to turn safe mode off.
{code}
> haadmin failover command does not provide enough detail for when target NN is
> not ready to be active
> ----------------------------------------------------------------------------------------------------
>
> Key: HDFS-3071
> URL: https://issues.apache.org/jira/browse/HDFS-3071
> Project: Hadoop HDFS
> Issue Type: Improvement
> Components: ha
> Affects Versions: 0.24.0
> Reporter: Philip Zeyliger
> Assignee: Todd Lipcon
> Attachments: hdfs-3071.txt
>
>
> When running the failover command, you can get an error message like the
> following:
> {quote}
> $ hdfs --config $(pwd) haadmin -failover namenode2 namenode1
> Failover failed: xxx.yyy/1.2.3.4:8020 is not ready to become active
> {quote}
> Unfortunately, the error message doesn't describe why that node isn't ready
> to be active. In my case, the target namenode's logs don't indicate anything
> either. It turned out that the issue was "Safe mode is ON.Resources are low
> on NN. Safe mode must be turned off manually.", but ideally the user would be
> told that at the time of the failover.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators:
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira