See <https://builds.apache.org/job/Hadoop-Hdfs-trunk/1525/changes>

Changes:

[acmurthy] Added ASL2 header to releasenotes.html.

[acmurthy] HADOOP-9350. Moving to appropriate section in CHANGES.txt

[acmurthy] Preparing for hadoop-2.1.1-beta release.

[acmurthy] HADOOP-9944. Fix RpcRequestHeaderProto.callId to be sint32 rather 
than uint32 since ipc.Client.CONNECTION_CONTEXT_CALL_ID is signed (i.e. -3). 
Contributed by Arun C. Murthy.

[szetszwo] HDFS-5207. In BlockPlacementPolicy.chooseTarget(..), change the 
writer and the excludedNodes parameter types respectively to Node and Set.  
Contributed by Junping Du

[cdouglas] YARN-1184. ClassCastException during preemption enforcement.

[vinodkv] YARN-1001. Added a web-service to get statistics about per 
application-type per state for consumption by downstream projects. Contributed 
by Zhijie Shen.

[cnauroth] HADOOP-9935. Revert from trunk.

[bikas] YARN-1027. Implement RMHAProtocolService (Karthik Kambatla via bikas)

[acmurthy] MAPREDUCE-5493. Cleanup in-memory & on-disk segments to prevent leak 
on shuffle completion. Contributed by Jason Lowe.

[acmurthy] YARN-1170. YARN & MapReduce proto definitions fixed to specify 
protobuf package as hadoop.yarn and hadoop.mapreduce respectively. Contibuted 
by Binglin Chang.

[tucu] MAPREDUCE-5379. Include token tracking ids in jobconf. (kkambatl via 
tucu)

[tucu] HADOOP-9962. in order to avoid dependency divergence within Hadoop 
itself lets enable DependencyConvergence. (rvs via tucu)

[tucu] HADOOP-9961. versions of a few transitive dependencies diverged between 
hadoop subprojects. (rvs via tucu)

------------------------------------------
[...truncated 9413 lines...]
+ /home/jenkins/tools/maven/latest/bin/mvn clean verify checkstyle:checkstyle 
findbugs:findbugs -Pdist -Pnative -Dtar -Pdocs -fae
[INFO] Scanning for projects...
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Build Order:
[INFO] 
[INFO] Apache Hadoop HDFS
[INFO] Apache Hadoop HttpFS
[INFO] Apache Hadoop HDFS BookKeeper Journal
[INFO] Apache Hadoop HDFS-NFS
[INFO] Apache Hadoop HDFS Project
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building Apache Hadoop HDFS 3.0.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-clean-plugin:2.4.1:clean (default-clean) @ hadoop-hdfs ---
[INFO] Deleting 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target>
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (create-testdirs) @ hadoop-hdfs ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test-dir>
    [mkdir] Created dir: 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data>
[INFO] Executed tasks
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (create-jsp-generated-sources-directory) 
@ hadoop-hdfs ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/generated-sources/java>
[INFO] Executed tasks
[INFO] 
[INFO] --- jspc-maven-plugin:2.0-alpha-3:compile (hdfs) @ hadoop-hdfs ---
[WARNING] Compiled JSPs will not be added to the project and web.xml will not 
be modified, either because includeInProject is set to false or because the 
project's packaging is not 'war'.
Created dir: 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/classes>
[INFO] Compiling 8 JSP source files to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/generated-sources/java>
log4j:WARN No appenders could be found for logger (org.apache.jasper.JspC).
log4j:WARN Please initialize the log4j system properly.
WARN: The method class 
org.apache.commons.logging.impl.SLF4JLogFactory#release() was invoked.
WARN: Please see http://www.slf4j.org/codes.html for an explanation.
[INFO] Compiled completed in 0:00:00.249
[INFO] 
[INFO] --- jspc-maven-plugin:2.0-alpha-3:compile (secondary) @ hadoop-hdfs ---
[WARNING] Compiled JSPs will not be added to the project and web.xml will not 
be modified, either because includeInProject is set to false or because the 
project's packaging is not 'war'.
[INFO] Compiling 1 JSP source file to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/generated-sources/java>
WARN: The method class 
org.apache.commons.logging.impl.SLF4JLogFactory#release() was invoked.
WARN: Please see http://www.slf4j.org/codes.html for an explanation.
[INFO] Compiled completed in 0:00:00.017
[INFO] 
[INFO] --- jspc-maven-plugin:2.0-alpha-3:compile (journal) @ hadoop-hdfs ---
[WARNING] Compiled JSPs will not be added to the project and web.xml will not 
be modified, either because includeInProject is set to false or because the 
project's packaging is not 'war'.
[INFO] Compiling 1 JSP source file to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/generated-sources/java>
WARN: The method class 
org.apache.commons.logging.impl.SLF4JLogFactory#release() was invoked.
WARN: Please see http://www.slf4j.org/codes.html for an explanation.
[INFO] Compiled completed in 0:00:00.015
[INFO] 
[INFO] --- jspc-maven-plugin:2.0-alpha-3:compile (datanode) @ hadoop-hdfs ---
[WARNING] Compiled JSPs will not be added to the project and web.xml will not 
be modified, either because includeInProject is set to false or because the 
project's packaging is not 'war'.
[INFO] Compiling 4 JSP source files to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/generated-sources/java>
WARN: The method class 
org.apache.commons.logging.impl.SLF4JLogFactory#release() was invoked.
WARN: Please see http://www.slf4j.org/codes.html for an explanation.
[INFO] Compiled completed in 0:00:00.025
[INFO] 
[INFO] --- build-helper-maven-plugin:1.5:add-source 
(add-jsp-generated-sources-directory) @ hadoop-hdfs ---
[INFO] Source directory: 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/generated-sources/java>
 added.
[INFO] 
[INFO] --- hadoop-maven-plugins:3.0.0-SNAPSHOT:protoc (compile-protoc) @ 
hadoop-hdfs ---
[INFO] 
[INFO] --- hadoop-maven-plugins:3.0.0-SNAPSHOT:protoc (compile-protoc-datanode) 
@ hadoop-hdfs ---
[INFO] 
[INFO] --- hadoop-maven-plugins:3.0.0-SNAPSHOT:protoc (compile-protoc-namenode) 
@ hadoop-hdfs ---
[INFO] 
[INFO] --- hadoop-maven-plugins:3.0.0-SNAPSHOT:protoc (compile-protoc-qjournal) 
@ hadoop-hdfs ---
[INFO] 
[INFO] --- maven-resources-plugin:2.2:resources (default-resources) @ 
hadoop-hdfs ---
[INFO] Using default encoding to copy filtered resources.
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hadoop-hdfs 
---
[INFO] Compiling 521 source files to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/classes>
[WARNING] 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/XmlEditsVisitor.java>:[32,48]
 com.sun.org.apache.xml.internal.serialize.OutputFormat is Sun proprietary API 
and may be removed in a future release
[WARNING] 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/XmlEditsVisitor.java>:[33,48]
 com.sun.org.apache.xml.internal.serialize.XMLSerializer is Sun proprietary API 
and may be removed in a future release
[WARNING] 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/XmlEditsVisitor.java>:[55,4]
 com.sun.org.apache.xml.internal.serialize.OutputFormat is Sun proprietary API 
and may be removed in a future release
[WARNING] 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/XmlEditsVisitor.java>:[55,33]
 com.sun.org.apache.xml.internal.serialize.OutputFormat is Sun proprietary API 
and may be removed in a future release
[WARNING] 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/XmlEditsVisitor.java>:[59,4]
 com.sun.org.apache.xml.internal.serialize.XMLSerializer is Sun proprietary API 
and may be removed in a future release
[WARNING] 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineEditsViewer/XmlEditsVisitor.java>:[59,35]
 com.sun.org.apache.xml.internal.serialize.XMLSerializer is Sun proprietary API 
and may be removed in a future release
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (create-web-xmls) @ hadoop-hdfs ---
[INFO] Executing tasks

main:
     [copy] Copying 1 file to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/webapps/hdfs/WEB-INF>
     [copy] Copying 1 file to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/webapps/secondary/WEB-INF>
     [copy] Copying 1 file to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/webapps/datanode/WEB-INF>
     [copy] Copying 1 file to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/webapps/journal/WEB-INF>
     [copy] Copying 9 files to 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/webapps>
[INFO] Executed tasks
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (make) @ hadoop-hdfs ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/native>
     [exec] -- The C compiler identification is GNU
     [exec] -- The CXX compiler identification is GNU
     [exec] -- Check for working C compiler: /usr/bin/gcc
     [exec] CMake Error at 
/usr/share/cmake-2.8/Modules/CMakeTestCCompiler.cmake:50 (MESSAGE):
     [exec]   The C compiler "/usr/bin/gcc" is not able to compile a simple 
test program.
     [exec] 
     [exec]   It fails with the following output:
     [exec] 
     [exec]    Change Dir: 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/native/CMakeFiles/CMakeTmp>
     [exec] 
     [exec]   
     [exec] 
     [exec]   Run Build Command:/usr/bin/make "cmTryCompileExec/fast"
     [exec] 
     [exec]   /usr/bin/make -f CMakeFiles/cmTryCompileExec.dir/build.make
     [exec]   CMakeFiles/cmTryCompileExec.dir/build
     [exec] 
     [exec]   make[1]: Entering directory
     [exec]   
`<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/ta-->
 Check for working C compiler: /usr/bin/gcc -- broken
     [exec] -- Configuring incomplete, errors occurred!
     [exec] rget/native/CMakeFiles/CMakeTmp'
     [exec] 
     [exec] 
     [exec]   /usr/bin/cmake -E cmake_progress_report
     [exec]   
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/native/CMakeFiles/CMakeTmp/CMakeFiles>
     [exec]   1
     [exec] 
     [exec]   Building C object 
CMakeFiles/cmTryCompileExec.dir/testCCompiler.c.o
     [exec] 
     [exec]   /usr/bin/gcc -o CMakeFiles/cmTryCompileExec.dir/testCCompiler.c.o 
-c
     [exec]   
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/native/CMakeFiles/CMakeTmp/testCCompiler.c>
     [exec] 
     [exec] 
     [exec]   Linking C executable cmTryCompileExec
     [exec] 
     [exec]   /usr/bin/cmake -E cmake_link_script
     [exec]   CMakeFiles/cmTryCompileExec.dir/link.txt --verbose=1
     [exec] 
     [exec]   /usr/bin/gcc CMakeFiles/cmTryCompileExec.dir/testCCompiler.c.o -o
     [exec]   cmTryCompileExec -rdynamic
     [exec] 
     [exec]   gcc: vfork: Resource temporarily unavailable
     [exec] 
     [exec]   make[1]: *** [cmTryCompileExec] Error 1
     [exec] 
     [exec]   make[1]: Leaving directory
     [exec]   
`<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs/target/native/CMakeFiles/CMakeTmp'>
     [exec] 
     [exec] 
     [exec]   make: *** [cmTryCompileExec/fast] Error 2
     [exec] 
     [exec]   
     [exec] 
     [exec]   
     [exec] 
     [exec]   CMake will not be able to correctly generate this project.
     [exec] Call Stack (most recent call first):
     [exec]   
     [exec] 
     [exec] 
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Skipping Apache Hadoop HttpFS
[INFO] This project has been banned from the build due to previous failures.
[INFO] ------------------------------------------------------------------------
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Skipping Apache Hadoop HDFS BookKeeper Journal
[INFO] This project has been banned from the build due to previous failures.
[INFO] ------------------------------------------------------------------------
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Skipping Apache Hadoop HDFS-NFS
[INFO] This project has been banned from the build due to previous failures.
[INFO] ------------------------------------------------------------------------
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building Apache Hadoop HDFS Project 3.0.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[WARNING] The POM for org.eclipse.m2e:lifecycle-mapping:jar:1.0.0 is missing, 
no dependency information available
[WARNING] Failed to retrieve plugin descriptor for 
org.eclipse.m2e:lifecycle-mapping:1.0.0: Plugin 
org.eclipse.m2e:lifecycle-mapping:1.0.0 or one of its dependencies could not be 
resolved: Failed to read artifact descriptor for 
org.eclipse.m2e:lifecycle-mapping:jar:1.0.0
[INFO] 
[INFO] --- maven-clean-plugin:2.4.1:clean (default-clean) @ hadoop-hdfs-project 
---
[INFO] Deleting 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/target>
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (create-testdirs) @ hadoop-hdfs-project 
---
[INFO] Executing tasks

main:
    [mkdir] Created dir: 
<https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/target/test-dir>
[INFO] Executed tasks
[INFO] 
[INFO] --- maven-source-plugin:2.1.2:jar-no-fork (hadoop-java-sources) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- maven-source-plugin:2.1.2:test-jar-no-fork (hadoop-java-sources) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- maven-enforcer-plugin:1.3.1:enforce (dist-enforce) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- maven-site-plugin:3.0:attach-descriptor (attach-descriptor) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- maven-javadoc-plugin:2.8.1:jar (module-javadocs) @ 
hadoop-hdfs-project ---
[INFO] Not executing Javadoc as the project is not a Java classpath-capable 
package
[INFO] 
[INFO] --- maven-enforcer-plugin:1.3.1:enforce (depcheck) @ hadoop-hdfs-project 
---
[INFO] 
[INFO] --- maven-checkstyle-plugin:2.6:checkstyle (default-cli) @ 
hadoop-hdfs-project ---
[INFO] 
[INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default-cli) @ 
hadoop-hdfs-project ---
[INFO] ****** FindBugsMojo execute *******
[INFO] canGenerate is false
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Hadoop HDFS ................................ FAILURE [39.498s]
[INFO] Apache Hadoop HttpFS .............................. SKIPPED
[INFO] Apache Hadoop HDFS BookKeeper Journal ............. SKIPPED
[INFO] Apache Hadoop HDFS-NFS ............................ SKIPPED
[INFO] Apache Hadoop HDFS Project ........................ SUCCESS [1.692s]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 42.050s
[INFO] Finished at: Tue Sep 17 11:34:13 UTC 2013
[INFO] Final Memory: 45M/432M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-antrun-plugin:1.6:run (make) on project 
hadoop-hdfs: An Ant BuildException has occured: exec returned: 1 -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
Build step 'Execute shell' marked build as failure
Archiving artifacts
Updating HDFS-5207
Updating HADOOP-9962
Updating HADOOP-9944
Updating YARN-1170
Updating HADOOP-9961
Updating MAPREDUCE-5379
Updating YARN-1027
Updating HADOOP-9935
Updating YARN-1184
Updating MAPREDUCE-5493
Updating YARN-1001
Updating HADOOP-9350

Reply via email to