[ 
https://issues.apache.org/jira/browse/HADOOP-1135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12482955
 ] 

dhruba borthakur commented on HADOOP-1135:
------------------------------------------

I agree that this could cause data loss. 

> A block report processing may incorrectly cause the namenode to delete blocks 
> ------------------------------------------------------------------------------
>
>                 Key: HADOOP-1135
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1135
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>            Reporter: dhruba borthakur
>         Assigned To: dhruba borthakur
>         Attachments: blockReportInvalidateBlock2.patch
>
>
> When a block report arrives at the namenode, the namenode goes through all 
> the blocks on that datanode. If a block is not valid it is marked for 
> deletion. The blocks-to-be-deleted are sent to the datanode as a response to 
> the next heartbeat RPC. The namenode sends only 100 blocks-to-be-deleted at a 
> time. This was introduced as part of hadoop-994. The bug is that if the 
> number of blocks-to-be-deleted exceeds 100, then that namenode marks all the 
> remaining blocks in the block report for deletion.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to