See https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2806/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE 
###########################
[...truncated 33014 lines...]
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 75.239 sec - in 
org.apache.hadoop.mapred.TestReduceFetchFromPartialMem
Running org.apache.hadoop.mapred.TestSpecialCharactersInOutputPath
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.394 sec - in 
org.apache.hadoop.mapred.TestSpecialCharactersInOutputPath
Running org.apache.hadoop.mapred.TestSequenceFileAsBinaryOutputFormat
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.99 sec - in 
org.apache.hadoop.mapred.TestSequenceFileAsBinaryOutputFormat

Results :

Failed tests: 
  TestNetworkedJob.testNetworkedJob:174 expected:<[[Sat Jan 09 03:04:09 +0000 
2016] Application is Activated, waiting for resources to be assigned for AM.  
Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = 
<memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's 
Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> 
but was:<[]>

Tests in error: 
  TestFixedLengthInputFormat.testFormat:81->runRandomTests:328 » NoClassDefFound
  
TestFixedLengthInputFormat.testPartialRecordUncompressedIn:190->runPartialRecordTest:410->readSplit:371
 » NoClassDefFound
  TestMiniMRWithDFSWithDistinctUsers.setUp:97 » YarnRuntime 
java.lang.NoClassDef...
  TestMiniMRWithDFSWithDistinctUsers.setUp:78 » IllegalState

Tests run: 527, Failures: 1, Errors: 4, Skipped: 11

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Hadoop MapReduce Client .................... SUCCESS [  2.925 s]
[INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:54 min]
[INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 29.033 s]
[INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [  4.867 s]
[INFO] Apache Hadoop MapReduce App ....................... SUCCESS [09:34 min]
[INFO] Apache Hadoop MapReduce HistoryServer ............. SUCCESS [05:55 min]
[INFO] Apache Hadoop MapReduce JobClient ................. FAILURE [  01:41 h]
[INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED
[INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED
[INFO] Apache Hadoop MapReduce Examples .................. SKIPPED
[INFO] Apache Hadoop MapReduce ........................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 02:00 h
[INFO] Finished at: 2016-01-09T03:11:38+00:00
[INFO] Final Memory: 34M/748M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on 
project hadoop-mapreduce-client-jobclient: There are test failures.
[ERROR] 
[ERROR] Please refer to 
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire-reports
 for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hadoop-mapreduce-client-jobclient
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any



###################################################################################
############################## FAILED TESTS (if any) 
##############################
5 tests failed.
FAILED:  org.apache.hadoop.mapred.TestFixedLengthInputFormat.testFormat

Error Message:
org/apache/hadoop/fs/ContentSummary$Builder

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/fs/ContentSummary$Builder
        at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
        at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
        at java.security.AccessController.doPrivileged(Native Method)
        at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
        at 
org.apache.hadoop.fs.FileSystem.getContentSummary(FileSystem.java:1480)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSDataBoundedInputStream.getFileLength(ChecksumFileSystem.java:279)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSDataBoundedInputStream.seek(ChecksumFileSystem.java:322)
        at 
org.apache.hadoop.mapreduce.lib.input.FixedLengthRecordReader.initialize(FixedLengthRecordReader.java:108)
        at 
org.apache.hadoop.mapred.FixedLengthRecordReader.<init>(FixedLengthRecordReader.java:47)
        at 
org.apache.hadoop.mapred.FixedLengthInputFormat.getRecordReader(FixedLengthInputFormat.java:87)
        at 
org.apache.hadoop.mapred.TestFixedLengthInputFormat.runRandomTests(TestFixedLengthInputFormat.java:328)
        at 
org.apache.hadoop.mapred.TestFixedLengthInputFormat.testFormat(TestFixedLengthInputFormat.java:81)


FAILED:  
org.apache.hadoop.mapred.TestFixedLengthInputFormat.testPartialRecordUncompressedIn

Error Message:
org/apache/hadoop/fs/ContentSummary$Builder

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/fs/ContentSummary$Builder
        at 
org.apache.hadoop.fs.FileSystem.getContentSummary(FileSystem.java:1480)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSDataBoundedInputStream.getFileLength(ChecksumFileSystem.java:279)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSDataBoundedInputStream.seek(ChecksumFileSystem.java:322)
        at 
org.apache.hadoop.mapreduce.lib.input.FixedLengthRecordReader.initialize(FixedLengthRecordReader.java:108)
        at 
org.apache.hadoop.mapred.FixedLengthRecordReader.<init>(FixedLengthRecordReader.java:47)
        at 
org.apache.hadoop.mapred.FixedLengthInputFormat.getRecordReader(FixedLengthInputFormat.java:87)
        at 
org.apache.hadoop.mapred.TestFixedLengthInputFormat.readSplit(TestFixedLengthInputFormat.java:371)
        at 
org.apache.hadoop.mapred.TestFixedLengthInputFormat.runPartialRecordTest(TestFixedLengthInputFormat.java:410)
        at 
org.apache.hadoop.mapred.TestFixedLengthInputFormat.testPartialRecordUncompressedIn(TestFixedLengthInputFormat.java:190)


FAILED:  
org.apache.hadoop.mapred.TestMiniMRWithDFSWithDistinctUsers.testMultipleSpills

Error Message:
java.lang.NoClassDefFoundError: 
org/apache/hadoop/hdfs/protocol/proto/ClientNamenodeProtocolProtos$GetListingRequestProto$Builder

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: 
java.lang.NoClassDefFoundError: 
org/apache/hadoop/hdfs/protocol/proto/ClientNamenodeProtocolProtos$GetListingRequestProto$Builder
        at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
        at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
        at java.security.AccessController.doPrivileged(Native Method)
        at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
        at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetListingRequestProto.newBuilder(ClientNamenodeProtocolProtos.java:29103)
        at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:542)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:255)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:103)
        at com.sun.proxy.$Proxy25.getListing(Unknown Source)
        at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1575)
        at org.apache.hadoop.fs.Hdfs$DirListingIterator.<init>(Hdfs.java:211)
        at org.apache.hadoop.fs.Hdfs$DirListingIterator.<init>(Hdfs.java:198)
        at org.apache.hadoop.fs.Hdfs$2.<init>(Hdfs.java:180)
        at org.apache.hadoop.fs.Hdfs.listStatusIterator(Hdfs.java:180)
        at org.apache.hadoop.fs.FileContext$22.next(FileContext.java:1492)
        at org.apache.hadoop.fs.FileContext$22.next(FileContext.java:1487)
        at org.apache.hadoop.fs.FSLinkResolver.resolve(FSLinkResolver.java:90)
        at org.apache.hadoop.fs.FileContext.listStatus(FileContext.java:1494)
        at 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils.localGlobber(JobHistoryUtils.java:456)
        at 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils.localGlobber(JobHistoryUtils.java:444)
        at 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils.localGlobber(JobHistoryUtils.java:439)
        at 
org.apache.hadoop.mapreduce.v2.hs.HistoryFileManager.findTimestampedDirectories(HistoryFileManager.java:809)
        at 
org.apache.hadoop.mapreduce.v2.hs.HistoryFileManager.initExisting(HistoryFileManager.java:703)
        at 
org.apache.hadoop.mapreduce.v2.hs.JobHistory.serviceInit(JobHistory.java:97)
        at 
org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
        at 
org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
        at 
org.apache.hadoop.mapreduce.v2.hs.JobHistoryServer.serviceInit(JobHistoryServer.java:152)
        at 
org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
        at 
org.apache.hadoop.mapreduce.v2.MiniMRYarnCluster$JobHistoryServerWrapper.serviceStart(MiniMRYarnCluster.java:232)
        at 
org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
        at 
org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:120)
        at 
org.apache.hadoop.mapreduce.v2.MiniMRYarnCluster.serviceStart(MiniMRYarnCluster.java:191)
        at 
org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
        at 
org.apache.hadoop.mapred.MiniMRClientClusterFactory.create(MiniMRClientClusterFactory.java:80)
        at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:187)
        at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:175)
        at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:167)
        at 
org.apache.hadoop.mapred.TestMiniMRWithDFSWithDistinctUsers.setUp(TestMiniMRWithDFSWithDistinctUsers.java:97)


FAILED:  
org.apache.hadoop.mapred.TestMiniMRWithDFSWithDistinctUsers.testDistinctUsers

Error Message:
null

Stack Trace:
java.lang.IllegalStateException: null
        at 
com.google.common.base.Preconditions.checkState(Preconditions.java:129)
        at org.apache.hadoop.ipc.Client.setCallIdAndRetryCount(Client.java:123)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:100)
        at com.sun.proxy.$Proxy25.getDatanodeReport(Unknown Source)
        at org.apache.hadoop.hdfs.DFSClient.datanodeReport(DFSClient.java:2035)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.waitActive(MiniDFSCluster.java:2447)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.waitActive(MiniDFSCluster.java:2490)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1632)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:844)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:482)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:441)
        at 
org.apache.hadoop.mapred.TestMiniMRWithDFSWithDistinctUsers.setUp(TestMiniMRWithDFSWithDistinctUsers.java:78)


FAILED:  org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob

Error Message:
expected:<[[Sat Jan 09 03:04:09 +0000 2016] Application is Activated, waiting 
for resources to be assigned for AM.  Details : AM Partition = 
<DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's 
Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's 
Absolute max capacity = 100.0 % ; ]> but was:<[]>

Stack Trace:
org.junit.ComparisonFailure: expected:<[[Sat Jan 09 03:04:09 +0000 2016] 
Application is Activated, waiting for resources to be assigned for AM.  Details 
: AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, 
vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used 
capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
        at org.junit.Assert.assertEquals(Assert.java:115)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob(TestNetworkedJob.java:174)


Reply via email to