[ https://issues.apache.org/jira/browse/HDFS-9373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15062843#comment-15062843 ]
Hudson commented on HDFS-9373: ------------------------------ FAILURE: Integrated in Hadoop-trunk-Commit #8987 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/8987/]) HDFS-9373. Erasure coding: friendly log information for write operations (zhz: rev 5104077e1f431ad3675d0b1c5c3cf53936902d8e) * hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSStripedOutputStream.java * hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt > Erasure coding: friendly log information for write operations with some > failed streamers > ---------------------------------------------------------------------------------------- > > Key: HDFS-9373 > URL: https://issues.apache.org/jira/browse/HDFS-9373 > Project: Hadoop HDFS > Issue Type: Sub-task > Components: erasure-coding > Affects Versions: 3.0.0 > Reporter: Li Bo > Assignee: Li Bo > Fix For: 3.0.0 > > Attachments: HDFS-9373-001.patch, HDFS-9373-002.patch, > HDFS-9373-003.patch > > > When not more than PARITY_NUM streamers fail for a block group, the client > may still succeed to write the data. But several exceptions are thrown to > user and user has to check the reasons. The friendly way is just inform user > that some streamers fail when writing a block group. It’s not necessary to > show the details of exceptions because a small number of stream failures is > not vital to the client writing. > When only DATA_NUM streamers succeed, the block group is in a high risk > because the corrupt of any block will cause all the six blocks' data lost. We > should give obvious warning to user when this occurs. -- This message was sent by Atlassian JIRA (v6.3.4#6332)