TestAppend2#testComplexAppend failed on "Too many open files"
-------------------------------------------------------------

                 Key: HDFS-690
                 URL: https://issues.apache.org/jira/browse/HDFS-690
             Project: Hadoop HDFS
          Issue Type: Bug
          Components: test
    Affects Versions: 0.21.0
            Reporter: Hairong Kuang
            Priority: Blocker
             Fix For: 0.21.0


the append write failed on "Too many open files":
Some bytes were failed to append to a file on the following error:
java.io.IOException: Cannot run program "stat": java.io.IOException: error=24, 
Too many open files
        at java.lang.ProcessBuilder.start(ProcessBuilder.java:459)
        at java.lang.Runtime.exec(Runtime.java:593)
        at java.lang.Runtime.exec(Runtime.java:466)
        at 
org.apache.hadoop.fs.FileUtil$HardLink.getLinkCount(FileUtil.java:644)
        at 
org.apache.hadoop.hdfs.server.datanode.ReplicaInfo.unlinkBlock(ReplicaInfo.java:205)
        at 
org.apache.hadoop.hdfs.server.datanode.FSDataset.append(FSDataset.java:1075)
        at 
org.apache.hadoop.hdfs.server.datanode.FSDataset.append(FSDataset.java:1058)
        at 
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.<init>(BlockReceiver.java:110)
        at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.opWriteBlock(DataXceiver.java:258)
        at 
org.apache.hadoop.hdfs.protocol.DataTransferProtocol$Receiver.opWriteBlock(DataTransferProtocol.java:382)
        at 
org.apache.hadoop.hdfs.protocol.DataTransferProtocol$Receiver.processOp(DataTransferProtocol.java:323)
        at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:111)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to