Walter Su created HDFS-8881: ------------------------------- Summary: Erasure Coding: internal blocks got missed and got over-replicated at the same time Key: HDFS-8881 URL: https://issues.apache.org/jira/browse/HDFS-8881 Project: Hadoop HDFS Issue Type: Sub-task Reporter: Walter Su Assignee: Walter Su
We know the Repl checking depends on {{BlockManager#countNodes()}}, but countNodes() has limitation for striped blockGroup. *One* missing internal block will be catched by Repl checking, and handled by ReplicationMonitor. *One* over-replicated internal block will be catched by Repl checking, and handled by processOverReplicatedBlocks. *One* missing internal block and *two* over-replicated internal blocks *at the same time* will be catched by Repl checking, and handled by processOverReplicatedBlocks, later by ReplicationMonitor. *One* missing internal block and *One* over-replicated internal block *at the same time* will *NOT* be catched by Repl checking. "at the same time" means one missing internal block can't be recovered, and one internal block got over-replicated anyway. For example: scenario A: step 1. block #0 and #1 are reported missing. 2. a new #1 got recovered. 3. the old #1 come back, and the recovery work for #0 failed. scenario B: 1. An DN decommissioned/dead which has #1. 2. block #0 is reported missing. 3. The DN has #1 recommisioned, and the recovery work for #0 failed. In the end, the blockGroup has \[1, 1, 2, 3, 4, 5, 6, 7, 8\], assume 6+3 schema. Client always needs to decode #0 if the blockGroup doesn't get handled. -- This message was sent by Atlassian JIRA (v6.3.4#6332)