[ 
https://issues.apache.org/jira/browse/HDFS-7570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14276256#comment-14276256
 ] 

Hudson commented on HDFS-7570:
------------------------------

FAILURE: Integrated in Hadoop-trunk-Commit #6855 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/6855/])
HDFS-7570. SecondaryNameNode need twice memory when calling 
reloadFromImageFile. Contributed by zhaoyunjiong. (cnauroth: rev 
85aec75ce53445e1abf840076d2e10f1e3c6d69b)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlocksMap.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> DataXceiver could leak FileDescriptor
> -------------------------------------
>
>                 Key: HDFS-7570
>                 URL: https://issues.apache.org/jira/browse/HDFS-7570
>             Project: Hadoop HDFS
>          Issue Type: Bug
>            Reporter: Juan Yu
>
> DataXceiver doesn't close inputstream all the time, There could be FD leakage 
> and overtime cause FDs exceed limit.
> {code}
> finally {
>       if (LOG.isDebugEnabled()) {
>         LOG.debug(datanode.getDisplayName() + ":Number of active connections 
> is: "
>             + datanode.getXceiverCount());
>       }
>       updateCurrentThreadName("Cleaning up");
>       if (peer != null) {
>         dataXceiverServer.closePeer(peer);
>         IOUtils.closeStream(in);
>       }
>     }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to