[ https://issues.apache.org/jira/browse/HDFS-2932?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Srikanth Upputuri resolved HDFS-2932. ------------------------------------- Resolution: Duplicate Fix Version/s: (was: 0.24.0) Closed as duplicate of HDFS-3493. > Under replicated block after the pipeline recovery. > --------------------------------------------------- > > Key: HDFS-2932 > URL: https://issues.apache.org/jira/browse/HDFS-2932 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode > Affects Versions: 0.24.0 > Reporter: J.Andreina > Assignee: Srikanth Upputuri > > Started 1NN,DN1,DN2,DN3 in the same machine. > Written a huge file of size 2 Gb > while the write for the block-id-1005 is in progress bruought down DN3. > after the pipeline recovery happened.Block stamp changed into block_id_1006 > in DN1,Dn2. > after the write is over.DN3 is brought up and fsck command is issued. > the following mess is displayed as follows > "block-id_1006 is underreplicatede.Target replicas is 3 but found 2 replicas". -- This message was sent by Atlassian JIRA (v6.3.4#6332)