Lantao Jin created HDFS-11285:
---------------------------------
Summary: Dead DataNodes keep a long time in (Dead,
DECOMMISSION_INPROGRESS), and never transition to (Dead, DECOMMISSIONED)
Key: HDFS-11285
URL: https://issues.apache.org/jira/browse/HDFS-11285
Project: Hadoop HDFS
Issue Type: Bug
Affects Versions: 2.7.1
Reporter: Lantao Jin
We have seen the use case of decommissioning DataNodes that are already dead or
unresponsive, and not expected to rejoin the cluster. In a large cluster, we
met more than 100 nodes were dead, decommissioning and their {panel} Under
replicated blocks {panel} {panel} Blocks with no live replicas {panel} were all
ZERO. Actually It has been fixed in
[HDFS-7374|https://issues.apache.org/jira/browse/HDFS-7374]. After that, we can
refreshNode twice to eliminate this case. But, seems this patch missed after
refactor[HDFS-7411|https://issues.apache.org/jira/browse/HDFS-7411]. We are
using a Hadoop version based 2.7.1 and only below operations can transition the
status from {panel} Dead, DECOMMISSION_INPROGRESS {panel} to {panel} Dead,
DECOMMISSIONED {panel}:
# Retire it from hdfs-exclude
# refreshNodes
# Re-add it to hdfs-exclude
# refreshNodes
So, why the code removed after refactor in the new DecommissionManager?
{code:java}
if (!node.isAlive) {
LOG.info("Dead node " + node + " is decommissioned immediately.");
node.setDecommissioned();
{code}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]