See https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/142/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE 
###########################
[...truncated 7075 lines...]
[INFO] 
[INFO] --- maven-source-plugin:2.3:jar-no-fork (hadoop-java-sources) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- maven-source-plugin:2.3:test-jar-no-fork (hadoop-java-sources) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- maven-enforcer-plugin:1.3.1:enforce (dist-enforce) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- maven-site-plugin:3.4:attach-descriptor (attach-descriptor) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- maven-javadoc-plugin:2.8.1:jar (module-javadocs) @ 
hadoop-hdfs-project ---
[INFO] Not executing Javadoc as the project is not a Java classpath-capable 
package
[INFO] 
[INFO] --- maven-enforcer-plugin:1.3.1:enforce (depcheck) @ hadoop-hdfs-project 
---
[INFO] 
[INFO] --- maven-checkstyle-plugin:2.12.1:checkstyle (default-cli) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- findbugs-maven-plugin:3.0.0:findbugs (default-cli) @ 
hadoop-hdfs-project ---
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Hadoop HDFS ................................ FAILURE [  03:07 h]
[INFO] Apache Hadoop HttpFS .............................. SKIPPED
[INFO] Apache Hadoop HDFS BookKeeper Journal ............. SKIPPED
[INFO] Apache Hadoop HDFS-NFS ............................ SKIPPED
[INFO] Apache Hadoop HDFS Project ........................ SUCCESS [  1.806 s]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 03:07 h
[INFO] Finished at: 2015-04-02T14:56:44+00:00
[INFO] Final Memory: 50M/269M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on 
project hadoop-hdfs: There was a timeout or other error in the fork -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
Build step 'Execute shell' marked build as failure
Archiving artifacts
Recording test results
Updating HDFS-8008
Updating HDFS-8026
Updating HDFS-8009
Updating YARN-3430
Updating HDFS-7978
Updating HDFS-7916
Updating HADOOP-11731
Updating HDFS-8001
Updating YARN-3248
Updating HDFS-7922
Updating YARN-3425
Updating HDFS-8036
Updating HADOOP-11787
Updating HADOOP-11757
Sending e-mails to: hdfs-dev@hadoop.apache.org
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) 
##############################
2 tests failed.
REGRESSION:  
org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting.testDataNodeReconfigureWithVolumeFailures

Error Message:
java.util.zip.ZipException: invalid stored block lengths

Stack Trace:
java.lang.RuntimeException: java.util.zip.ZipException: invalid stored block 
lengths
        at java.util.zip.InflaterInputStream.read(InflaterInputStream.java:164)
        at java.util.zip.InflaterInputStream.read(InflaterInputStream.java:122)
        at java.io.FilterInputStream.read(FilterInputStream.java:83)
        at 
org.apache.xerces.impl.XMLEntityManager$RewindableInputStream.read(Unknown 
Source)
        at org.apache.xerces.impl.XMLEntityManager.setupCurrentEntity(Unknown 
Source)
        at 
org.apache.xerces.impl.XMLVersionDetector.determineDocVersion(Unknown Source)
        at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
        at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
        at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
        at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)
        at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)
        at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2501)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2489)
        at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2560)
        at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2513)
        at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2424)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1160)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1132)
        at org.apache.hadoop.conf.Configuration.setLong(Configuration.java:1377)
        at 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting.initCluster(TestDataNodeVolumeFailureReporting.java:556)
        at 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting.setUp(TestDataNodeVolumeFailureReporting.java:84)


REGRESSION:  
org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting.testVolFailureStatsPreservedOnNNRestart

Error Message:
java.util.zip.ZipException: invalid stored block lengths

Stack Trace:
java.lang.RuntimeException: java.util.zip.ZipException: invalid stored block 
lengths
        at java.util.zip.InflaterInputStream.read(InflaterInputStream.java:164)
        at java.util.zip.InflaterInputStream.read(InflaterInputStream.java:122)
        at java.io.FilterInputStream.read(FilterInputStream.java:83)
        at 
org.apache.xerces.impl.XMLEntityManager$RewindableInputStream.read(Unknown 
Source)
        at org.apache.xerces.impl.XMLEntityManager.setupCurrentEntity(Unknown 
Source)
        at 
org.apache.xerces.impl.XMLVersionDetector.determineDocVersion(Unknown Source)
        at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
        at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
        at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
        at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)
        at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)
        at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2501)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2489)
        at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2560)
        at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2513)
        at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2424)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1160)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1132)
        at org.apache.hadoop.conf.Configuration.setLong(Configuration.java:1377)
        at 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting.initCluster(TestDataNodeVolumeFailureReporting.java:556)
        at 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting.setUp(TestDataNodeVolumeFailureReporting.java:84)


Reply via email to