[ 
https://issues.apache.org/jira/browse/HDFS-4906?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13684006#comment-13684006
 ] 

Hadoop QA commented on HDFS-4906:
---------------------------------

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12587936/HDFS-4906.patch
  against trunk revision .

    {color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

    {color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

    {color:red}-1 javac{color:red}.  The patch appears to cause the build to 
fail.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4523//console

This message is automatically generated.
                
> HDFS Output streams should not accept writes after being closed
> ---------------------------------------------------------------
>
>                 Key: HDFS-4906
>                 URL: https://issues.apache.org/jira/browse/HDFS-4906
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: hdfs-client
>    Affects Versions: 2.0.5-alpha
>            Reporter: Aaron T. Myers
>            Assignee: Aaron T. Myers
>         Attachments: HDFS-4906.patch
>
>
> Currently if one closes an OutputStream obtained from FileSystem#create and 
> then calls write(...) on that closed stream, the write will appear to succeed 
> without error though no data will be written to HDFS. A subsequent call to 
> close will also silently appear to succeed. We should make it so that 
> attempts to write to closed streams fails fast.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to