[ https://issues.apache.org/jira/browse/HDFS-4360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13544615#comment-13544615 ]
Hadoop QA commented on HDFS-4360: --------------------------------- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12563403/HDFS-4360.patch against trunk revision . {color:red}-1 patch{color}. The patch command could not apply the patch. Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/3744//console This message is automatically generated. > multiple BlockFixer should be supported in order to improve scalability and > reduce too much work on single BlockFixer > --------------------------------------------------------------------------------------------------------------------- > > Key: HDFS-4360 > URL: https://issues.apache.org/jira/browse/HDFS-4360 > Project: Hadoop HDFS > Issue Type: Improvement > Components: contrib/raid > Affects Versions: 0.22.0 > Reporter: Jun Jin > Labels: patch > Attachments: HDFS-4360.patch > > > current implementation can only run single BlockFixer since the fsck (in > RaidDFSUtil.getCorruptFiles) only check the whole DFS file system. multiple > BlockFixer will do the same thing and try to fix same file if multiple > BlockFixer launched. > the change/fix will be mainly in BlockFixer.java and > RaidDFSUtil.getCorruptFile(), to enable fsck to check the different paths > defined in separated Raid.xml for single RaidNode/BlockFixer -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira