Lantao Jin created HDFS-11285:
---------------------------------

             Summary: Dead DataNodes keep a long time in (Dead, 
DECOMMISSION_INPROGRESS), and never transition to (Dead, DECOMMISSIONED)
                 Key: HDFS-11285
                 URL: https://issues.apache.org/jira/browse/HDFS-11285
             Project: Hadoop HDFS
          Issue Type: Bug
    Affects Versions: 2.7.1
            Reporter: Lantao Jin


We have seen the use case of decommissioning DataNodes that are already dead or 
unresponsive, and not expected to rejoin the cluster. In a large cluster, we 
met more than 100 nodes were dead, decommissioning and their {panel} Under 
replicated blocks {panel} {panel} Blocks with no live replicas {panel} were all 
ZERO. Actually It has been fixed in 
[HDFS-7374|https://issues.apache.org/jira/browse/HDFS-7374]. After that, we can 
refreshNode twice to eliminate this case. But, seems this patch missed after 
refactor[HDFS-7411|https://issues.apache.org/jira/browse/HDFS-7411]. We are 
using a Hadoop version based 2.7.1 and only below operations can transition the 
status from {panel} Dead, DECOMMISSION_INPROGRESS {panel} to {panel} Dead, 
DECOMMISSIONED {panel}:
# Retire it from hdfs-exclude
# refreshNodes
# Re-add it to hdfs-exclude
# refreshNodes

So, why the code removed after refactor in the new DecommissionManager?
{code:java}
if (!node.isAlive) {
  LOG.info("Dead node " + node + " is decommissioned immediately.");
  node.setDecommissioned();
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to