[ https://issues.apache.org/jira/browse/HDFS-11285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15907420#comment-15907420 ]
Lantao Jin edited comment on HDFS-11285 at 3/13/17 1:10 PM: ------------------------------------------------------------ Hi [~andrew.wang], I found there are lots of logs in NameNode like below: {code} 2017-03-13 03:59:52,620 INFO org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager: Block: blk_13651215184_1113964818077{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=2, replicas=[ReplicaUC[[DISK]DS-a74fff1e-dc86-4e60-8e69-9c9023a7fd3c:NORMAL:10.115.21.54:50010|RBW]]}, Expected Replicas: 3, live replicas: 0, corrupt replicas: 0, decommissioned replicas: 0, decommissioning replicas: 1, excess replicas: 0, Is Open File: true, Datanodes having this block: 10.115.21.54:50010 , Current Datanode: 10.115.21.54:50010, Is current datanode decommissioning: true {code} *But the block blk_13651215184 can't be found in that node 10.115.21.54.* Notice the UCState=COMMITTED, and this will cause decommission_inprogress never completed. was (Author: cltlfcjin): Hi [~andrew.wang], I found there are lots of logs in NameNode like below: {code} 2017-03-13 03:59:52,620 INFO org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager: Block: blk_13651215184_111 3964818077{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=2, replicas=[ReplicaUC[[DISK]DS-a74fff1e-dc86-4e60-8e69 -9c9023a7fd3c:NORMAL:10.115.21.54:50010|RBW]]}, Expected Replicas: 3, live replicas: 0, corrupt replicas: 0, decommissione d replicas: 0, decommissioning replicas: 1, excess replicas: 0, Is Open File: true, Datanodes having this block: 10.115.21 .54:50010 , Current Datanode: 10.115.21.54:50010, Is current datanode decommissioning: true {code} *But the block blk_13651215184 can't be found in that node 10.115.21.54.* Notice the UCState=COMMITTED, and this will cause decommission_inprogress never completed. > Dead DataNodes keep a long time in (Dead, DECOMMISSION_INPROGRESS), and never > transition to (Dead, DECOMMISSIONED) > ------------------------------------------------------------------------------------------------------------------ > > Key: HDFS-11285 > URL: https://issues.apache.org/jira/browse/HDFS-11285 > Project: Hadoop HDFS > Issue Type: Bug > Affects Versions: 2.7.1 > Reporter: Lantao Jin > Attachments: DecomStatus.png > > > We have seen the use case of decommissioning DataNodes that are already dead > or unresponsive, and not expected to rejoin the cluster. In a large cluster, > we met more than 100 nodes were dead, decommissioning and their {{Under > replicated blocks}} {{Blocks with no live replicas}} were all ZERO. Actually > It has been fixed in > [HDFS-7374|https://issues.apache.org/jira/browse/HDFS-7374]. After that, we > can refreshNode twice to eliminate this case. But, seems this patch missed > after refactor[HDFS-7411|https://issues.apache.org/jira/browse/HDFS-7411]. We > are using a Hadoop version based 2.7.1 and only below operations can > transition the status from {{Dead, DECOMMISSION_INPROGRESS}} to {{Dead, > DECOMMISSIONED}}: > # Retire it from hdfs-exclude > # refreshNodes > # Re-add it to hdfs-exclude > # refreshNodes > So, why the code removed after refactor in the new DecommissionManager? > {code:java} > if (!node.isAlive) { > LOG.info("Dead node " + node + " is decommissioned immediately."); > node.setDecommissioned(); > {code} -- This message was sent by Atlassian JIRA (v6.3.15#6346) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org