[ https://issues.apache.org/jira/browse/HDFS-729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12806459#action_12806459 ]
Raghu Angadi commented on HDFS-729: ----------------------------------- As I understand, list of files returned in one call will not have duplicates. 500 is a lot.. note that these are files with 'hard-corruption', ie, HDFS could not repair them. Once a cluster has so many corrupt files, I would think there would be a lot more urgent things to worry about than finding rest of the corrupt files. In practice, most likely reason for such a scenario would be a large number of datanodes go missing. > fsck option to list only corrupted files > ---------------------------------------- > > Key: HDFS-729 > URL: https://issues.apache.org/jira/browse/HDFS-729 > Project: Hadoop HDFS > Issue Type: Improvement > Reporter: dhruba borthakur > Assignee: dhruba borthakur > Attachments: badFiles.txt, badFiles2.txt, corruptFiles.txt > > > An option to fsck to list only corrupted files will be very helpful for > frequent monitoring. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.