HDFS should repair corrupted files
----------------------------------

                 Key: HADOOP-855
                 URL: https://issues.apache.org/jira/browse/HADOOP-855
             Project: Hadoop
          Issue Type: Bug
          Components: dfs
            Reporter: Wendy Chien
         Assigned To: Wendy Chien


While reading if we discover a mismatch between a block and checksum, we want 
to report this back to the namenode to delete the corrupted block or crc.

To implement this, we need to do the following:
DFSInputStream
1. move DFSInputStream out of DFSClient
2. add member variable to keep track of current datanode (the chosen node)

DistributedFileSystem
1. change reportChecksumFailure parameter crc from int to FSInputStream (needed 
to be able to delete it). 
2. determine specific block and datanode from DFSInputStream passed to 
reportChecksumFailure  
3. call namenode to delete block/crc vis DFSClient

ClientProtocol
1. add method to ask namenode to delete certain blocks on specifc datanode.

Namenode
1. add ability to delete certain blocks on specific datanode

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators: 
https://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to