Walter Su created HDFS-8881:
-------------------------------

             Summary: Erasure Coding: internal blocks got missed and got 
over-replicated at the same time
                 Key: HDFS-8881
                 URL: https://issues.apache.org/jira/browse/HDFS-8881
             Project: Hadoop HDFS
          Issue Type: Sub-task
            Reporter: Walter Su
            Assignee: Walter Su


We know the Repl checking depends on {{BlockManager#countNodes()}}, but 
countNodes() has limitation for striped blockGroup.

*One* missing internal block will be catched by Repl checking, and handled by 
ReplicationMonitor.
*One* over-replicated internal block will be catched by Repl checking, and 
handled by processOverReplicatedBlocks.
*One* missing internal block and *two* over-replicated internal blocks *at the 
same time* will be catched by Repl checking, and handled by 
processOverReplicatedBlocks, later by ReplicationMonitor.
*One* missing internal block and *One* over-replicated internal block *at the 
same time* will *NOT* be catched by Repl checking.

"at the same time" means one missing internal block can't be recovered, and one 
internal block got over-replicated anyway. For example:

scenario A:
step 1. block #0 and #1 are reported missing.
2. a new #1 got recovered.
3. the old #1 come back, and the recovery work for #0 failed.

scenario B:
1. An DN decommissioned/dead which has #1.
2. block #0 is reported missing.
3. The DN has #1 recommisioned, and the recovery work for #0 failed.

In the end, the blockGroup has \[1, 1, 2, 3, 4, 5, 6, 7, 8\], assume 6+3 
schema. Client always needs to decode #0 if the blockGroup doesn't get handled.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to