[ https://issues.apache.org/jira/browse/HDFS-1480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13054217#comment-13054217 ]
Todd Lipcon commented on HDFS-1480: ----------------------------------- There's definitely a bug still in trunk, here. Here's my analysis (in this case with repl=3): - Cluster has 2 racks, A and B, each with three nodes. - Block initially written on A1, B1, B2 - Admin decommissions two of these nodes (let's say A1 and B1 but it doesn't matter) - in {{computeReplicationWorkForBlock}}, it calls {{chooseSourceDatanode}} which populates {{containingNodes}} with all three nodes regardless of decom status - this is then passed through as {{chosenNodes}} in the call to {{chooseTarget}} - maxNodesPerRack is then assigned (5-1)/2 + 2 = 4 - {{chooseTarget}} initializes {{result}} with these same three nodes - {numOfResults}} is thus 3, and it calls {{chooseRandom}} rather than basing its decision on local/remote racks - {{chooseRandom}} is free to pick any two nodes since maxNodesPerRack is 4 Will attach a failing unit test momentarily > All replicas for a block with repl=2 end up in same rack > -------------------------------------------------------- > > Key: HDFS-1480 > URL: https://issues.apache.org/jira/browse/HDFS-1480 > Project: Hadoop HDFS > Issue Type: Bug > Components: name-node > Affects Versions: 0.20.2 > Reporter: T Meyarivan > > It appears that all replicas of a block can end up in the same rack. The > likelihood of such replicas seems to be directly related to decommissioning > of nodes. > Post rolling OS upgrade (decommission 3-10% of nodes, re-install etc, add > them back) of a running cluster, all replicas of about 0.16% of blocks ended > up in the same rack. > Hadoop Namenode UI etc doesn't seem to know about such incorrectly replicated > blocks. "hadoop fsck .." does report that the blocks must be replicated on > additional racks. > Looking at ReplicationTargetChooser.java, following seem suspect: > snippet-01: > {code} > int maxNodesPerRack = > (totalNumOfReplicas-1)/clusterMap.getNumOfRacks()+2; > {code} > snippet-02: > {code} > case 2: > if (clusterMap.isOnSameRack(results.get(0), results.get(1))) { > chooseRemoteRack(1, results.get(0), excludedNodes, > blocksize, maxNodesPerRack, results); > } else if (newBlock){ > chooseLocalRack(results.get(1), excludedNodes, blocksize, > maxNodesPerRack, results); > } else { > chooseLocalRack(writer, excludedNodes, blocksize, > maxNodesPerRack, results); > } > if (--numOfReplicas == 0) { > break; > } > {code} > snippet-03: > {code} > do { > DatanodeDescriptor[] selectedNodes = > chooseRandom(1, nodes, excludedNodes); > if (selectedNodes.length == 0) { > throw new NotEnoughReplicasException( > "Not able to place enough > replicas"); > } > result = (DatanodeDescriptor)(selectedNodes[0]); > } while(!isGoodTarget(result, blocksize, maxNodesPerRack, results)); > {code} -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira