[jira] [Updated] (HDFS-9580) TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected number of invalidate blocks.
[ https://issues.apache.org/jira/browse/HDFS-9580?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Lowe updated HDFS-9580: - Fix Version/s: (was: 3.0.0-alpha1) 2.8.0 Thanks, [~jojochuang]! I committed this to branch-2 and branch-2.8 as well. > TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected > number of invalidate blocks. > -- > > Key: HDFS-9580 > URL: https://issues.apache.org/jira/browse/HDFS-9580 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode, namenode, test >Affects Versions: 3.0.0-alpha1 > Environment: Jenkins >Reporter: Wei-Chiu Chuang >Assignee: Wei-Chiu Chuang > Fix For: 2.8.0 > > Attachments: HDFS-9580.001.patch > > > The failure appeared in the trunk jenkins job. > https://builds.apache.org/job/Hadoop-Hdfs-trunk/2646/ > {noformat} > Error Message > Expected invalidate blocks to be the number of DNs expected:<3> but was:<2> > Stacktrace > java.lang.AssertionError: Expected invalidate blocks to be the number of DNs > expected:<3> but was:<2> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotEquals(Assert.java:743) > at org.junit.Assert.assertEquals(Assert.java:118) > at org.junit.Assert.assertEquals(Assert.java:555) > at > org.apache.hadoop.hdfs.server.blockmanagement.TestComputeInvalidateWork.testDatanodeReRegistration(TestComputeInvalidateWork.java:160) > {noformat} > I think there could be a race condition between creating a file and shutting > down data nodes, which failed the test. > {noformat} > 2015-12-19 07:11:02,765 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=LAST_IN_PIPELINE, downstreams=0:[]] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=LAST_IN_PIPELINE, downstreams=0:[] terminating > 2015-12-19 07:11:02,768 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace > (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:45655, dest: > /127.0.0.1:54890, bytes: 134217728, op: HDFS_WRITE, cliID: > DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: > 6a13ec05-e1c1-4086-8a4d-d5a09636afcd, blockid: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: > 954174423 > 2015-12-19 07:11:02,768 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE terminating > 2015-12-19 07:11:02,772 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace > (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:33252, dest: > /127.0.0.1:54426, bytes: 134217728, op: HDFS_WRITE, cliID: > DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: > d81751db-02a9-48fe-b697-77623048784b, blockid: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: > 957463510 > 2015-12-19 07:11:02,772 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE terminating > 2015-12-19 07:11:02,782 [IPC Server handler 4 on 36404] INFO > blockmanagement.BlockManager > (BlockManager.java:checkBlocksProperlyReplicated(3871)) - BLOCK* > blk_1073741825_1001 is not COMPLETE (ucState = COMMITTED, replication# = 0 < > minimum = 1) in file /testRR > 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO > namenode.EditLogFileOutputStream > (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush > 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO > namenode.EditLogFileOutputStream > (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush > 2015-12-19 07:11:03,190 [IPC Server handler 8 on 36404] INFO > hdfs.StateChange (FSNamesystem.java:completeFile(2557)) - DIR* completeFile: > /testRR is closed by DFSClient_NONMAPREDUCE_147911011_935 > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-9580) TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected number of invalidate blocks.
[ https://issues.apache.org/jira/browse/HDFS-9580?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhe Zhang updated HDFS-9580: Resolution: Fixed Hadoop Flags: Reviewed Fix Version/s: 3.0.0 Status: Resolved (was: Patch Available) Thanks Wei-Chiu. I just committed the patch to trunk. > TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected > number of invalidate blocks. > -- > > Key: HDFS-9580 > URL: https://issues.apache.org/jira/browse/HDFS-9580 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode, namenode, test >Affects Versions: 3.0.0 > Environment: Jenkins >Reporter: Wei-Chiu Chuang >Assignee: Wei-Chiu Chuang > Fix For: 3.0.0 > > Attachments: HDFS-9580.001.patch > > > The failure appeared in the trunk jenkins job. > https://builds.apache.org/job/Hadoop-Hdfs-trunk/2646/ > {noformat} > Error Message > Expected invalidate blocks to be the number of DNs expected:<3> but was:<2> > Stacktrace > java.lang.AssertionError: Expected invalidate blocks to be the number of DNs > expected:<3> but was:<2> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotEquals(Assert.java:743) > at org.junit.Assert.assertEquals(Assert.java:118) > at org.junit.Assert.assertEquals(Assert.java:555) > at > org.apache.hadoop.hdfs.server.blockmanagement.TestComputeInvalidateWork.testDatanodeReRegistration(TestComputeInvalidateWork.java:160) > {noformat} > I think there could be a race condition between creating a file and shutting > down data nodes, which failed the test. > {noformat} > 2015-12-19 07:11:02,765 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=LAST_IN_PIPELINE, downstreams=0:[]] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=LAST_IN_PIPELINE, downstreams=0:[] terminating > 2015-12-19 07:11:02,768 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace > (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:45655, dest: > /127.0.0.1:54890, bytes: 134217728, op: HDFS_WRITE, cliID: > DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: > 6a13ec05-e1c1-4086-8a4d-d5a09636afcd, blockid: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: > 954174423 > 2015-12-19 07:11:02,768 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE terminating > 2015-12-19 07:11:02,772 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace > (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:33252, dest: > /127.0.0.1:54426, bytes: 134217728, op: HDFS_WRITE, cliID: > DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: > d81751db-02a9-48fe-b697-77623048784b, blockid: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: > 957463510 > 2015-12-19 07:11:02,772 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE terminating > 2015-12-19 07:11:02,782 [IPC Server handler 4 on 36404] INFO > blockmanagement.BlockManager > (BlockManager.java:checkBlocksProperlyReplicated(3871)) - BLOCK* > blk_1073741825_1001 is not COMPLETE (ucState = COMMITTED, replication# = 0 < > minimum = 1) in file /testRR > 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO > namenode.EditLogFileOutputStream > (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush > 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO > namenode.EditLogFileOutputStream > (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush > 2015-12-19 07:11:03,190 [IPC Server handler 8 on 36404] INFO > hdfs.StateChange (FSNamesystem.java:completeFile(2557)) - DIR* completeFile: > /testRR is closed by DFSClient_NONMAPREDUCE_147911011_935 > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HDFS-9580) TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected number of invalidate blocks.
[ https://issues.apache.org/jira/browse/HDFS-9580?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Wei-Chiu Chuang updated HDFS-9580: -- Status: Patch Available (was: Open) > TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected > number of invalidate blocks. > -- > > Key: HDFS-9580 > URL: https://issues.apache.org/jira/browse/HDFS-9580 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode, namenode, test >Affects Versions: 3.0.0 > Environment: Jenkins >Reporter: Wei-Chiu Chuang >Assignee: Wei-Chiu Chuang > Attachments: HDFS-9580.001.patch > > > The failure appeared in the trunk jenkins job. > https://builds.apache.org/job/Hadoop-Hdfs-trunk/2646/ > {noformat} > Error Message > Expected invalidate blocks to be the number of DNs expected:<3> but was:<2> > Stacktrace > java.lang.AssertionError: Expected invalidate blocks to be the number of DNs > expected:<3> but was:<2> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotEquals(Assert.java:743) > at org.junit.Assert.assertEquals(Assert.java:118) > at org.junit.Assert.assertEquals(Assert.java:555) > at > org.apache.hadoop.hdfs.server.blockmanagement.TestComputeInvalidateWork.testDatanodeReRegistration(TestComputeInvalidateWork.java:160) > {noformat} > I think there could be a race condition between creating a file and shutting > down data nodes, which failed the test. > {noformat} > 2015-12-19 07:11:02,765 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=LAST_IN_PIPELINE, downstreams=0:[]] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=LAST_IN_PIPELINE, downstreams=0:[] terminating > 2015-12-19 07:11:02,768 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace > (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:45655, dest: > /127.0.0.1:54890, bytes: 134217728, op: HDFS_WRITE, cliID: > DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: > 6a13ec05-e1c1-4086-8a4d-d5a09636afcd, blockid: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: > 954174423 > 2015-12-19 07:11:02,768 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE terminating > 2015-12-19 07:11:02,772 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace > (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:33252, dest: > /127.0.0.1:54426, bytes: 134217728, op: HDFS_WRITE, cliID: > DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: > d81751db-02a9-48fe-b697-77623048784b, blockid: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: > 957463510 > 2015-12-19 07:11:02,772 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE terminating > 2015-12-19 07:11:02,782 [IPC Server handler 4 on 36404] INFO > blockmanagement.BlockManager > (BlockManager.java:checkBlocksProperlyReplicated(3871)) - BLOCK* > blk_1073741825_1001 is not COMPLETE (ucState = COMMITTED, replication# = 0 < > minimum = 1) in file /testRR > 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO > namenode.EditLogFileOutputStream > (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush > 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO > namenode.EditLogFileOutputStream > (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush > 2015-12-19 07:11:03,190 [IPC Server handler 8 on 36404] INFO > hdfs.StateChange (FSNamesystem.java:completeFile(2557)) - DIR* completeFile: > /testRR is closed by DFSClient_NONMAPREDUCE_147911011_935 > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HDFS-9580) TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected number of invalidate blocks.
[ https://issues.apache.org/jira/browse/HDFS-9580?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Wei-Chiu Chuang updated HDFS-9580: -- Attachment: HDFS-9580.001.patch Rev01 patch. resolve race condition by ensuring the file is properly replicated before shutting down data nodes. > TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected > number of invalidate blocks. > -- > > Key: HDFS-9580 > URL: https://issues.apache.org/jira/browse/HDFS-9580 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode, namenode, test >Affects Versions: 3.0.0 > Environment: Jenkins >Reporter: Wei-Chiu Chuang >Assignee: Wei-Chiu Chuang > Attachments: HDFS-9580.001.patch > > > The failure appeared in the trunk jenkins job. > https://builds.apache.org/job/Hadoop-Hdfs-trunk/2646/ > {noformat} > Error Message > Expected invalidate blocks to be the number of DNs expected:<3> but was:<2> > Stacktrace > java.lang.AssertionError: Expected invalidate blocks to be the number of DNs > expected:<3> but was:<2> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotEquals(Assert.java:743) > at org.junit.Assert.assertEquals(Assert.java:118) > at org.junit.Assert.assertEquals(Assert.java:555) > at > org.apache.hadoop.hdfs.server.blockmanagement.TestComputeInvalidateWork.testDatanodeReRegistration(TestComputeInvalidateWork.java:160) > {noformat} > I think there could be a race condition between creating a file and shutting > down data nodes, which failed the test. > {noformat} > 2015-12-19 07:11:02,765 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=LAST_IN_PIPELINE, downstreams=0:[]] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=LAST_IN_PIPELINE, downstreams=0:[] terminating > 2015-12-19 07:11:02,768 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace > (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:45655, dest: > /127.0.0.1:54890, bytes: 134217728, op: HDFS_WRITE, cliID: > DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: > 6a13ec05-e1c1-4086-8a4d-d5a09636afcd, blockid: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: > 954174423 > 2015-12-19 07:11:02,768 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE terminating > 2015-12-19 07:11:02,772 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace > (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:33252, dest: > /127.0.0.1:54426, bytes: 134217728, op: HDFS_WRITE, cliID: > DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: > d81751db-02a9-48fe-b697-77623048784b, blockid: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: > 957463510 > 2015-12-19 07:11:02,772 [PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode > (BlockReceiver.java:run(1404)) - PacketResponder: > BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, > type=HAS_DOWNSTREAM_IN_PIPELINE terminating > 2015-12-19 07:11:02,782 [IPC Server handler 4 on 36404] INFO > blockmanagement.BlockManager > (BlockManager.java:checkBlocksProperlyReplicated(3871)) - BLOCK* > blk_1073741825_1001 is not COMPLETE (ucState = COMMITTED, replication# = 0 < > minimum = 1) in file /testRR > 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO > namenode.EditLogFileOutputStream > (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush > 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO > namenode.EditLogFileOutputStream > (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush > 2015-12-19 07:11:03,190 [IPC Server handler 8 on 36404] INFO > hdfs.StateChange (FSNamesystem.java:completeFile(2557)) - DIR* completeFile: > /testRR is closed by DFSClient_NONMAPREDUCE_147911011_935 > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HDFS-9580) TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected number of invalidate blocks.
[ https://issues.apache.org/jira/browse/HDFS-9580?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Wei-Chiu Chuang updated HDFS-9580: -- Description: The failure appeared in the trunk jenkins job. https://builds.apache.org/job/Hadoop-Hdfs-trunk/2646/ {noformat} Error Message Expected invalidate blocks to be the number of DNs expected:<3> but was:<2> Stacktrace java.lang.AssertionError: Expected invalidate blocks to be the number of DNs expected:<3> but was:<2> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:555) at org.apache.hadoop.hdfs.server.blockmanagement.TestComputeInvalidateWork.testDatanodeReRegistration(TestComputeInvalidateWork.java:160) {noformat} I think there could be a race condition between creating a file and shutting down data nodes, which failed the test. {noformat} 2015-12-19 07:11:02,765 [PacketResponder: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, type=LAST_IN_PIPELINE, downstreams=0:[]] INFO datanode.DataNode (BlockReceiver.java:run(1404)) - PacketResponder: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, type=LAST_IN_PIPELINE, downstreams=0:[] terminating 2015-12-19 07:11:02,768 [PacketResponder: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:45655, dest: /127.0.0.1:54890, bytes: 134217728, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: 6a13ec05-e1c1-4086-8a4d-d5a09636afcd, blockid: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: 954174423 2015-12-19 07:11:02,768 [PacketResponder: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode (BlockReceiver.java:run(1404)) - PacketResponder: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, type=HAS_DOWNSTREAM_IN_PIPELINE terminating 2015-12-19 07:11:02,772 [PacketResponder: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, type=HAS_DOWNSTREAM_IN_PIPELINE] INFO DataNode.clienttrace (BlockReceiver.java:finalizeBlock(1431)) - src: /127.0.0.1:33252, dest: /127.0.0.1:54426, bytes: 134217728, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_147911011_935, offset: 0, srvID: d81751db-02a9-48fe-b697-77623048784b, blockid: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, duration: 957463510 2015-12-19 07:11:02,772 [PacketResponder: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, type=HAS_DOWNSTREAM_IN_PIPELINE] INFO datanode.DataNode (BlockReceiver.java:run(1404)) - PacketResponder: BP-1551077294-67.195.81.149-1450509060247:blk_1073741825_1001, type=HAS_DOWNSTREAM_IN_PIPELINE terminating 2015-12-19 07:11:02,782 [IPC Server handler 4 on 36404] INFO blockmanagement.BlockManager (BlockManager.java:checkBlocksProperlyReplicated(3871)) - BLOCK* blk_1073741825_1001 is not COMPLETE (ucState = COMMITTED, replication# = 0 < minimum = 1) in file /testRR 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO namenode.EditLogFileOutputStream (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush 2015-12-19 07:11:02,783 [IPC Server handler 4 on 36404] INFO namenode.EditLogFileOutputStream (EditLogFileOutputStream.java:flushAndSync(200)) - Nothing to flush 2015-12-19 07:11:03,190 [IPC Server handler 8 on 36404] INFO hdfs.StateChange (FSNamesystem.java:completeFile(2557)) - DIR* completeFile: /testRR is closed by DFSClient_NONMAPREDUCE_147911011_935 {noformat} was: The failure appeared in the trunk jenkins job. https://builds.apache.org/job/Hadoop-Hdfs-trunk/2646/ {noformat} Error Message Expected invalidate blocks to be the number of DNs expected:<3> but was:<2> Stacktrace java.lang.AssertionError: Expected invalidate blocks to be the number of DNs expected:<3> but was:<2> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:555) at org.apache.hadoop.hdfs.server.blockmanagement.TestComputeInvalidateWork.testDatanodeReRegistration(TestComputeInvalidateWork.java:160) {noformat} I think there could be a race condition between creating a file and shutting down data nodes, which failed the test. > TestComputeInvalidateWork#testDatanodeReRegistration failed due to unexpected > number of invalidate blocks. > -- > > Key: HDFS-9580 > URL: https://issues.apache.org/jira/browse/HDFS-9580 > Proje