[ https://issues.apache.org/jira/browse/HDFS-12453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Arpit Agarwal updated HDFS-12453: --------------------------------- Description: TestDataNodeHotSwapVolumes fails occasionally with the following error (see comment). Ran it ~10 times locally and it passed every time. (was: TestDataNodeHotSwapVolumes fails occasionally with the following error. Ran it ~10 times locally and it passed every time. {code} Error Message Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[DatanodeInfoWithStorage[127.0.0.1:46950,DS-e653e0ec-e490-47d3-9cca-cdd6dff3964a,DISK], DatanodeInfoWithStorage[127.0.0.1:34486,DS-fd83f642-be1a-44ed-8ba5-6ad5a42e85fd,DISK]], original=[DatanodeInfoWithStorage[127.0.0.1:34486,DS-fd83f642-be1a-44ed-8ba5-6ad5a42e85fd,DISK], DatanodeInfoWithStorage[127.0.0.1:46950,DS-e653e0ec-e490-47d3-9cca-cdd6dff3964a,DISK]]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration. Stacktrace java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[DatanodeInfoWithStorage[127.0.0.1:46950,DS-e653e0ec-e490-47d3-9cca-cdd6dff3964a,DISK], DatanodeInfoWithStorage[127.0.0.1:34486,DS-fd83f642-be1a-44ed-8ba5-6ad5a42e85fd,DISK]], original=[DatanodeInfoWithStorage[127.0.0.1:34486,DS-fd83f642-be1a-44ed-8ba5-6ad5a42e85fd,DISK], DatanodeInfoWithStorage[127.0.0.1:46950,DS-e653e0ec-e490-47d3-9cca-cdd6dff3964a,DISK]]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration. at org.apache.hadoop.hdfs.DataStreamer.findNewDatanode(DataStreamer.java:1321) at org.apache.hadoop.hdfs.DataStreamer.addDatanode2ExistingPipeline(DataStreamer.java:1387) at org.apache.hadoop.hdfs.DataStreamer.handleDatanodeReplacement(DataStreamer.java:1586) at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1487) at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1469) at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1273) at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:684) {code}) > TestDataNodeHotSwapVolumes fails in trunk Jenkins runs > ------------------------------------------------------ > > Key: HDFS-12453 > URL: https://issues.apache.org/jira/browse/HDFS-12453 > Project: Hadoop HDFS > Issue Type: Bug > Components: test > Reporter: Arpit Agarwal > Priority: Critical > > TestDataNodeHotSwapVolumes fails occasionally with the following error (see > comment). Ran it ~10 times locally and it passed every time. -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org