[ https://issues.apache.org/jira/browse/HDFS-10720?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15408809#comment-15408809 ]
Rakesh R commented on HDFS-10720: --------------------------------- Thanks [~drankye] for the reviews. Attached new patch addressing the same. Also, I've added retires during {{getComputedDatanodeWork}} for safer side. Kindly review it again. > Fix intermittent test failure of TestDataNodeErasureCodingMetrics#testEcTasks > ----------------------------------------------------------------------------- > > Key: HDFS-10720 > URL: https://issues.apache.org/jira/browse/HDFS-10720 > Project: Hadoop HDFS > Issue Type: Bug > Reporter: Rakesh R > Assignee: Rakesh R > Attachments: HDFS-10720-00.patch, HDFS-10720-01.patch, > HDFS-10720-02.patch > > > The test is wrongly finding out the datanode to be corrupted from the block > locations. Instead of finding out a datanode which is used in the block > locations it is simply getting a datanode from the cluster, which may not be > a datanode present in the block locations. > {code} > byte[] indices = lastBlock.getBlockIndices(); > //corrupt the first block > DataNode toCorruptDn = cluster.getDataNodes().get(indices[0]); > {code} > For example, datanodes in the cluster.getDataNodes() array indexed like, > 0->Dn1, 1->Dn2, 2->Dn3, 3->Dn4, 4->Dn5, 5->Dn6, 6->Dn7, 7->Dn8, 8->Dn9, > 9->Dn10 > Assume the datanodes which are part of block location is => Dn2, Dn3, Dn4, > Dn5, Dn6, Dn7, Dn8, Dn9, Dn10. Now, in the failed scenario, it is getting the > corrupted datanode as cluster.getDataNodes().get(0) which will be Dn1 and > corruption of this datanode will not result in ECWork and is failing the > tests. > Ideally, the test should find a datanode from the block locations and corrupt > it, that will trigger ECWork. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org