Jira: https://issues.apache.org/jira/browse/OOZIE-2869 Build: https://builds.apache.org/job/PreCommit-OOZIE-Build/220/
################################################################################### ########################## LAST 60 LINES OF THE CONSOLE ########################### [...truncated 1.57 MB...] [INFO] --- maven-resources-plugin:2.5:testResources (default-testResources) @ oozie-zookeeper-security-tests --- [debug] execute contextualize [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] skip non existing resourceDirectory /home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/zookeeper-security-tests/src/test/resources [INFO] Copying 3 resources [INFO] [INFO] --- maven-compiler-plugin:2.3.2:testCompile (default-testCompile) @ oozie-zookeeper-security-tests --- [INFO] Nothing to compile - all classes are up to date [INFO] [INFO] --- maven-surefire-plugin:2.20.1:test (default-test) @ oozie-zookeeper-security-tests --- [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Oozie Main .................................. SUCCESS [ 0.578 s] [INFO] Apache Oozie Client ................................ SUCCESS [ 1.228 s] [INFO] Apache Oozie Share Lib Oozie ....................... SUCCESS [ 2.095 s] [INFO] Apache Oozie Share Lib HCatalog .................... SUCCESS [ 0.650 s] [INFO] Apache Oozie Share Lib Distcp ...................... SUCCESS [ 0.133 s] [INFO] Apache Oozie Core .................................. SUCCESS [ 5.797 s] [INFO] Apache Oozie Share Lib Streaming ................... SUCCESS [06:42 min] [INFO] Apache Oozie Share Lib Pig ......................... SUCCESS [04:05 min] [INFO] Apache Oozie Share Lib Hive ........................ SUCCESS [01:08 min] [INFO] Apache Oozie Share Lib Hive 2 ...................... SUCCESS [ 5.842 s] [INFO] Apache Oozie Share Lib Sqoop ....................... SUCCESS [ 2.898 s] [INFO] Apache Oozie Examples .............................. SUCCESS [ 5.057 s] [INFO] Apache Oozie Share Lib Spark ....................... SUCCESS [ 2.050 s] [INFO] Apache Oozie Share Lib ............................. SUCCESS [ 0.016 s] [INFO] Apache Oozie Docs .................................. SUCCESS [ 0.025 s] [INFO] Apache Oozie WebApp ................................ SUCCESS [ 2.374 s] [INFO] Apache Oozie Tools ................................. SUCCESS [ 1.138 s] [INFO] Apache Oozie MiniOozie ............................. SUCCESS [ 1.226 s] [INFO] Apache Oozie Server ................................ SUCCESS [ 2.849 s] [INFO] Apache Oozie Distro ................................ SUCCESS [ 1.088 s] [INFO] Apache Oozie ZooKeeper Security Tests .............. SUCCESS [ 1.658 s] [INFO] ------------------------------------------------------------------------ [INFO] BUILD SUCCESS [INFO] ------------------------------------------------------------------------ [INFO] Total time: 12:34 min [INFO] Finished at: 2017-11-14T19:44:29Z [INFO] Final Memory: 73M/1439M [INFO] ------------------------------------------------------------------------ bin/test-patch-20-tests: line 105: echo: write error: No space left on device Failure, check for details /home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/test-patch/tmp/test-patch-20-tests-post.out % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 3706k 0 5792 0 0 7292 0 0:08:40 --:--:-- 0:08:40 7292 curl: (23) Failed writing body (0 != 5792) Could not download jira-cli tool, thus no JIRA updating Build step 'Execute shell' marked build as failure [description-setter] Description set: OOZIE-2869 Archiving artifacts [Fast Archiver] Compressed 1.65 MB of artifacts by 39.7% relative to #216 Recording test results Email was triggered for: Failure - Any Sending email for trigger: Failure - Any ################################################################################### ############################## FAILED TESTS (if any) ############################## 61 tests failed. FAILED: org.apache.oozie.example.TestLocalOozieExample.testLocalOozieExampleEnd Error Message: Unable to initialize WebAppContext Stack Trace: java.io.IOException: Unable to initialize WebAppContext at org.apache.oozie.example.TestLocalOozieExample.setUp(TestLocalOozieExample.java:37) Caused by: java.io.IOException: No space left on device at org.apache.oozie.example.TestLocalOozieExample.setUp(TestLocalOozieExample.java:37) FAILED: org.apache.oozie.example.TestLocalOozieExample.testLocalOozieExampleKill Error Message: Unable to initialize WebAppContext Stack Trace: java.io.IOException: Unable to initialize WebAppContext at org.apache.oozie.example.TestLocalOozieExample.setUp(TestLocalOozieExample.java:37) Caused by: java.io.IOException: No space left on device at org.apache.oozie.example.TestLocalOozieExample.setUp(TestLocalOozieExample.java:37) FAILED: org.apache.oozie.action.hadoop.TestHiveActionExecutor.testSetupMethods Error Message: Timed out waiting for Mini HDFS Cluster to start Stack Trace: java.io.IOException: Timed out waiting for Mini HDFS Cluster to start FAILED: org.apache.oozie.action.hadoop.TestHiveActionExecutor.testActionConfLoadDefaultResources Error Message: Timed out waiting for Mini HDFS Cluster to start Stack Trace: java.io.IOException: Timed out waiting for Mini HDFS Cluster to start FAILED: org.apache.oozie.action.hadoop.TestHiveActionExecutor.testHiveAction Error Message: Unable to initialize WebAppContext Stack Trace: java.io.IOException: Unable to initialize WebAppContext Caused by: java.io.IOException: No space left on device FAILED: org.apache.oozie.action.hadoop.TestHiveMain.testJobIDPattern Error Message: Timed out waiting for Mini HDFS Cluster to start Stack Trace: java.io.IOException: Timed out waiting for Mini HDFS Cluster to start at org.apache.oozie.action.hadoop.TestHiveMain.setUp(TestHiveMain.java:44) FAILED: org.apache.oozie.action.hadoop.TestHiveMain.testMain Error Message: Timed out waiting for Mini HDFS Cluster to start Stack Trace: java.io.IOException: Timed out waiting for Mini HDFS Cluster to start at org.apache.oozie.action.hadoop.TestHiveMain.setUp(TestHiveMain.java:44) FAILED: org.apache.oozie.action.hadoop.TestHive2ActionExecutor.testSetupMethodsForQuery Error Message: Unable to initialize WebAppContext Stack Trace: java.io.IOException: Unable to initialize WebAppContext at org.apache.oozie.action.hadoop.TestHive2ActionExecutor.setUp(TestHive2ActionExecutor.java:56) Caused by: java.io.IOException: No space left on device at org.apache.oozie.action.hadoop.TestHive2ActionExecutor.setUp(TestHive2ActionExecutor.java:56) FAILED: org.apache.oozie.action.hadoop.TestHive2ActionExecutor.testHive2Action Error Message: /home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/hive2/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestHive2ActionExecutor/testHive2Action/dfe1b44f-ed70-4414-9d3a-2aee17b7d775/conf/hadoop-conf/hadoop-site.xml (No such file or directory) Stack Trace: java.io.FileNotFoundException: /home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/hive2/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestHive2ActionExecutor/testHive2Action/dfe1b44f-ed70-4414-9d3a-2aee17b7d775/conf/hadoop-conf/hadoop-site.xml (No such file or directory) at org.apache.oozie.action.hadoop.TestHive2ActionExecutor.setUp(TestHive2ActionExecutor.java:56) FAILED: org.apache.oozie.action.hadoop.TestHive2ActionExecutor.testHive2ActionFails Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/hive2/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestHive2ActionExecutor/testHive2ActionFails/3a9ab56b-2acf-4f13-bd66-d5430eb61bff] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/hive2/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestHive2ActionExecutor/testHive2ActionFails/3a9ab56b-2acf-4f13-bd66-d5430eb61bff] at org.apache.oozie.action.hadoop.TestHive2ActionExecutor.setUp(TestHive2ActionExecutor.java:56) FAILED: org.apache.oozie.action.hadoop.TestHive2ActionExecutor.testSetupMethodsForScript Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/hive2/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestHive2ActionExecutor/testSetupMethodsForScript/595216b5-9058-447d-b39d-8a1899f340ed] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/hive2/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestHive2ActionExecutor/testSetupMethodsForScript/595216b5-9058-447d-b39d-8a1899f340ed] at org.apache.oozie.action.hadoop.TestHive2ActionExecutor.setUp(TestHive2ActionExecutor.java:56) FAILED: org.apache.oozie.action.hadoop.TestPigActionExecutor.testPig Error Message: Timed out waiting for Mini HDFS Cluster to start Stack Trace: java.io.IOException: Timed out waiting for Mini HDFS Cluster to start at org.apache.oozie.action.hadoop.TestPigActionExecutor.setUp(TestPigActionExecutor.java:71) FAILED: org.apache.oozie.action.hadoop.TestPigActionExecutor.testExecutionStatsWithMaxStatsSizeLimit Error Message: YARN App state for app application_1510688362402_0001 expected:<FINISHED> but was:<ACCEPTED> Stack Trace: junit.framework.AssertionFailedError: YARN App state for app application_1510688362402_0001 expected:<FINISHED> but was:<ACCEPTED> at org.apache.oozie.action.hadoop.TestPigActionExecutor.testExecutionStatsWithMaxStatsSizeLimit(TestPigActionExecutor.java:271) FAILED: org.apache.oozie.action.hadoop.TestPigActionExecutor.testPigError Error Message: File /user/test/868bcf66-8fc3-41c1-ac93-615639ba503a/netty-3.6.2.Final.jar could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and 2 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: File /user/test/868bcf66-8fc3-41c1-ac93-615639ba503a/netty-3.6.2.Final.jar could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and 2 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) FAILED: org.apache.oozie.action.hadoop.TestPigActionExecutor.testActionConfLoadDefaultResources Error Message: File /user/test/9b97d57e-d34d-4090-bed5-2ff78be3b9e7/app/script.pig could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and 2 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: File /user/test/9b97d57e-d34d-4090-bed5-2ff78be3b9e7/app/script.pig could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and 2 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) FAILED: org.apache.oozie.action.hadoop.TestPigActionExecutor.testExecutionStatsWithRetrieveStatsFalse Error Message: File /user/test/90f5d1b3-7ceb-4785-8f90-a4c6033bd786/app/script.pig could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: File /user/test/90f5d1b3-7ceb-4785-8f90-a4c6033bd786/app/script.pig could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) FAILED: org.apache.oozie.action.hadoop.TestPigActionExecutor.testExternalChildIds Error Message: File /user/test/f457a10e-7377-4792-acf6-667bace2275e/app/script.pig could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: File /user/test/f457a10e-7377-4792-acf6-667bace2275e/app/script.pig could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) FAILED: org.apache.oozie.action.hadoop.TestPigActionExecutor.testExecutionStats Error Message: File /user/test/3d86af28-3dad-4c08-8b35-b4f9d5a4504e/app/script.pig could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: File /user/test/3d86af28-3dad-4c08-8b35-b4f9d5a4504e/app/script.pig could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) FAILED: org.apache.oozie.action.hadoop.TestPigActionExecutor.testUdfPig Error Message: File /user/test/fc0cbeb7-8767-4b3d-a250-79c685e5293e/udf.jar could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: File /user/test/fc0cbeb7-8767-4b3d-a250-79c685e5293e/udf.jar could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) FAILED: org.apache.oozie.action.hadoop.TestPigMain.testJobIDPattern Error Message: Cannot delete /user/test/b754fe31-1f53-4dc7-9da3-9dc249d06348. Name node is in safe mode. Resources are low on NN. Please add or free up more resources then turn off safe mode manually. NOTE: If you turn off safe mode before adding resources, the NN will immediately return to safe mode. Use "hdfs dfsadmin -safemode leave" to turn safe mode off. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:3967) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:3925) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3909) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:786) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:589) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: Cannot delete /user/test/b754fe31-1f53-4dc7-9da3-9dc249d06348. Name node is in safe mode. Resources are low on NN. Please add or free up more resources then turn off safe mode manually. NOTE: If you turn off safe mode before adding resources, the NN will immediately return to safe mode. Use "hdfs dfsadmin -safemode leave" to turn safe mode off. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:3967) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:3925) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3909) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:786) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:589) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) at org.apache.oozie.action.hadoop.TestPigMain.setUp(TestPigMain.java:47) FAILED: org.apache.oozie.action.hadoop.TestPigMain.testPig_withNullExternalID Error Message: Cannot delete /user/test/3fe696da-3175-4cd5-b923-b0ada0e7c504. Name node is in safe mode. Resources are low on NN. Please add or free up more resources then turn off safe mode manually. NOTE: If you turn off safe mode before adding resources, the NN will immediately return to safe mode. Use "hdfs dfsadmin -safemode leave" to turn safe mode off. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:3967) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:3925) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3909) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:786) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:589) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: Cannot delete /user/test/3fe696da-3175-4cd5-b923-b0ada0e7c504. Name node is in safe mode. Resources are low on NN. Please add or free up more resources then turn off safe mode manually. NOTE: If you turn off safe mode before adding resources, the NN will immediately return to safe mode. Use "hdfs dfsadmin -safemode leave" to turn safe mode off. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:3967) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:3925) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3909) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:786) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:589) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) at org.apache.oozie.action.hadoop.TestPigMain.setUp(TestPigMain.java:47) FAILED: org.apache.oozie.action.hadoop.TestPigMain.testEmbeddedPigWithinPython Error Message: Cannot delete /user/test/2221b27a-36aa-4985-b631-cd90f29c90d2. Name node is in safe mode. Resources are low on NN. Please add or free up more resources then turn off safe mode manually. NOTE: If you turn off safe mode before adding resources, the NN will immediately return to safe mode. Use "hdfs dfsadmin -safemode leave" to turn safe mode off. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:3967) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:3925) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3909) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:786) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:589) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: Cannot delete /user/test/2221b27a-36aa-4985-b631-cd90f29c90d2. Name node is in safe mode. Resources are low on NN. Please add or free up more resources then turn off safe mode manually. NOTE: If you turn off safe mode before adding resources, the NN will immediately return to safe mode. Use "hdfs dfsadmin -safemode leave" to turn safe mode off. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:3967) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:3925) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3909) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:786) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:589) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) at org.apache.oozie.action.hadoop.TestPigMain.setUp(TestPigMain.java:47) FAILED: org.apache.oozie.action.hadoop.TestPigMain.testPigScript Error Message: Cannot delete /user/test/66853dc6-7337-4d6b-83ff-cb197a006d1d. Name node is in safe mode. Resources are low on NN. Please add or free up more resources then turn off safe mode manually. NOTE: If you turn off safe mode before adding resources, the NN will immediately return to safe mode. Use "hdfs dfsadmin -safemode leave" to turn safe mode off. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:3967) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:3925) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3909) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:786) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:589) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: Cannot delete /user/test/66853dc6-7337-4d6b-83ff-cb197a006d1d. Name node is in safe mode. Resources are low on NN. Please add or free up more resources then turn off safe mode manually. NOTE: If you turn off safe mode before adding resources, the NN will immediately return to safe mode. Use "hdfs dfsadmin -safemode leave" to turn safe mode off. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:3967) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:3925) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3909) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:786) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:589) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) at org.apache.oozie.action.hadoop.TestPigMain.setUp(TestPigMain.java:47) FAILED: org.apache.oozie.action.hadoop.TestSparkActionExecutor.testSetupMethodsWithSparkConfiguration Error Message: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Stack Trace: java.net.ConnectException: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Caused by: java.net.ConnectException: Connection refused FAILED: org.apache.oozie.action.hadoop.TestSparkActionExecutor.testSetupMethods Error Message: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Stack Trace: java.net.ConnectException: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Caused by: java.net.ConnectException: Connection refused FAILED: org.apache.oozie.action.hadoop.TestSparkMain.testJobIDPattern Error Message: File /user/jenkins/target/MiniMRCluster_377820428-tmpDir/MRAppJar.jar could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and 2 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: File /user/jenkins/target/MiniMRCluster_377820428-tmpDir/MRAppJar.jar could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and 2 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) FAILED: org.apache.oozie.action.hadoop.TestSparkMain.testPatterns Error Message: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Stack Trace: java.net.ConnectException: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Caused by: java.net.ConnectException: Connection refused FAILED: org.apache.oozie.action.hadoop.TestSparkMain.testMain Error Message: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Stack Trace: java.net.ConnectException: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Caused by: java.net.ConnectException: Connection refused FAILED: org.apache.oozie.action.hadoop.TestSqoopActionExecutor.testSetupMethods Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSetupMethods/862724c7-3f69-47d9-92df-e97beb5687a3] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSetupMethods/862724c7-3f69-47d9-92df-e97beb5687a3] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduceWithUberJarEnabled Error Message: YARN App state for app application_1510687948111_0001 expected:<FINISHED> but was:<ACCEPTED> Stack Trace: junit.framework.AssertionFailedError: YARN App state for app application_1510687948111_0001 expected:<FINISHED> but was:<ACCEPTED> at org.apache.oozie.action.hadoop.TestMapReduceActionExecutor._testSubmit(TestMapReduceActionExecutor.java:400) at org.apache.oozie.action.hadoop.TestMapReduceActionExecutor._testMapReduceWithUberJar(TestMapReduceActionExecutor.java:727) at org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduceWithUberJarEnabled(TestMapReduceActionExecutor.java:815) FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testSetMapredJobName Error Message: YARN App state for app application_1510687948111_0002 expected:<FINISHED> but was:<ACCEPTED> Stack Trace: junit.framework.AssertionFailedError: YARN App state for app application_1510687948111_0002 expected:<FINISHED> but was:<ACCEPTED> at org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testSetMapredJobName(TestMapReduceActionExecutor.java:1144) FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduceWithUberJarDisabled Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithUberJarDisabled/2ba87b17-5f6f-4287-8e03-43664c11f4d8] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithUberJarDisabled/2ba87b17-5f6f-4287-8e03-43664c11f4d8] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testPipes Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testPipes/36db0ca4-f545-4a28-86a2-e667d11df8d1] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testPipes/36db0ca4-f545-4a28-86a2-e667d11df8d1] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testCommaSeparatedFilesAndArchives Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testCommaSeparatedFilesAndArchives/76ee0e80-b576-48d0-9d56-f00c5638491f] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testCommaSeparatedFilesAndArchives/76ee0e80-b576-48d0-9d56-f00c5638491f] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testEndWithoutConfiguration Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testEndWithoutConfiguration/2501aaa4-5554-47cd-8d11-7b9a808dad2a] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testEndWithoutConfiguration/2501aaa4-5554-47cd-8d11-7b9a808dad2a] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduceWithConfigClass Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithConfigClass/153cc883-0909-4414-bd61-d3e20baba80b] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithConfigClass/153cc883-0909-4414-bd61-d3e20baba80b] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduceWithCredentials Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithCredentials/bf093427-1f52-473b-a1b4-1e3f9ef0e00a] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithCredentials/bf093427-1f52-473b-a1b4-1e3f9ef0e00a] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testSetupMethods Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testSetupMethods/d66ade73-ea11-4ab5-b52f-ef50a0aeed1c] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testSetupMethods/d66ade73-ea11-4ab5-b52f-ef50a0aeed1c] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testJobNameSetForMapReduceChild Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testJobNameSetForMapReduceChild/55fe2e93-2951-4abb-8aa7-57a74d555455] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testJobNameSetForMapReduceChild/55fe2e93-2951-4abb-8aa7-57a74d555455] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testStreamingConfOverride Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testStreamingConfOverride/7fc211f6-87d1-4fee-9f53-2f176448d390] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testStreamingConfOverride/7fc211f6-87d1-4fee-9f53-2f176448d390] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testConfigDefaultPropsToAction Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testConfigDefaultPropsToAction/528afd7d-2cdd-4df6-a986-ca225dd3b71b] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testConfigDefaultPropsToAction/528afd7d-2cdd-4df6-a986-ca225dd3b71b] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduceWithConfigClassThrowException Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithConfigClassThrowException/99b7679d-d1b1-4340-8555-0affdf3ca9ee] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithConfigClassThrowException/99b7679d-d1b1-4340-8555-0affdf3ca9ee] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testStreaming Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testStreaming/db8fb176-bfff-4ce9-bb25-40f701e29c0d] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testStreaming/db8fb176-bfff-4ce9-bb25-40f701e29c0d] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testSetExecutionStats_when_user_has_specified_stats_write_TRUE Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testSetExecutionStats_when_user_has_specified_stats_write_TRUE/72b4b968-70c5-485a-8e21-5c2605ca3ff7] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testSetExecutionStats_when_user_has_specified_stats_write_TRUE/72b4b968-70c5-485a-8e21-5c2605ca3ff7] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testSetExecutionStats_when_user_has_specified_stats_write_FALSE Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testSetExecutionStats_when_user_has_specified_stats_write_FALSE/d713e740-58d8-49a3-a6f5-4eb22d1fccc3] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testSetExecutionStats_when_user_has_specified_stats_write_FALSE/d713e740-58d8-49a3-a6f5-4eb22d1fccc3] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduceWithConfigClassNotFound Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithConfigClassNotFound/311829fb-9709-49f7-aeb8-366c162b564d] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceWithConfigClassNotFound/311829fb-9709-49f7-aeb8-366c162b564d] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduce Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduce/6a5650f3-f190-4074-b6ec-eb3945753f8b] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduce/6a5650f3-f190-4074-b6ec-eb3945753f8b] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduceActionError Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceActionError/486296f9-c664-4aae-9d5b-8bb1b5e417dc] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceActionError/486296f9-c664-4aae-9d5b-8bb1b5e417dc] FAILED: org.apache.oozie.action.hadoop.TestMapReduceActionExecutor.testMapReduceActionKill Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceActionKill/ba54ffc3-fe55-449f-ae17-741e32f9fb01] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/streaming/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestMapReduceActionExecutor/testMapReduceActionKill/ba54ffc3-fe55-449f-ae17-741e32f9fb01] FAILED: org.apache.oozie.tools.TestOozieSharelibCLI.testOozieSharelibCLICreate Error Message: File /user/jenkins/target/MiniMRCluster_667895151-tmpDir/MRAppJar.jar could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and 2 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) Stack Trace: org.apache.hadoop.ipc.RemoteException: File /user/jenkins/target/MiniMRCluster_667895151-tmpDir/MRAppJar.jar could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and 2 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1549) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3200) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:641) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033) FAILED: org.apache.oozie.test.TestWorkflow.testParallelFsAndShellWorkflowCompletesSuccessfully Error Message: expected:<SUCCEEDED> but was:<RUNNING> Stack Trace: junit.framework.AssertionFailedError: expected:<SUCCEEDED> but was:<RUNNING> at org.apache.oozie.test.TestWorkflow.runWorkflowFromFile(TestWorkflow.java:167) at org.apache.oozie.test.TestWorkflow.testParallelFsAndShellWorkflowCompletesSuccessfully(TestWorkflow.java:117) FAILED: org.apache.oozie.test.TestWorkflowRetries.testParallelFsAndShellWorkflowCompletesSuccessfully Error Message: expected:<SUCCEEDED> but was:<RUNNING> Stack Trace: junit.framework.AssertionFailedError: expected:<SUCCEEDED> but was:<RUNNING> FAILED: org.apache.oozie.action.hadoop.TestPyspark.testPyspark Error Message: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Stack Trace: java.net.ConnectException: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Caused by: java.net.ConnectException: Connection refused FAILED: org.apache.oozie.action.hadoop.TestSparkActionExecutor.testSparkAction Error Message: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Stack Trace: java.net.ConnectException: Call From asf918.gq1.ygridcore.net/67.195.81.138 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused Caused by: java.net.ConnectException: Connection refused FAILED: org.apache.oozie.action.hadoop.TestSqoopActionExecutor.testSqoopActionWithRedundantArgsAndFreeFormQuery Error Message: Cannot create directory /home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/build/test/data/dfs/name1/current Stack Trace: java.io.IOException: Cannot create directory /home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/build/test/data/dfs/name1/current FAILED: org.apache.oozie.action.hadoop.TestSqoopActionExecutor.testSqoopActionWithArgsAndFreeFormQuery Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopActionWithArgsAndFreeFormQuery/64f39b9c-0c9d-4dd7-81b6-7943a54798f8] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopActionWithArgsAndFreeFormQuery/64f39b9c-0c9d-4dd7-81b6-7943a54798f8] FAILED: org.apache.oozie.action.hadoop.TestSqoopActionExecutor.testSqoopActionWithBadRedundantArgsAndFreeFormQuery Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopActionWithBadRedundantArgsAndFreeFormQuery/1a466781-2517-43c1-babf-7e0156737bf8] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopActionWithBadRedundantArgsAndFreeFormQuery/1a466781-2517-43c1-babf-7e0156737bf8] FAILED: org.apache.oozie.action.hadoop.TestSqoopActionExecutor.testSqoopActionWithBadCommand Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopActionWithBadCommand/626a6841-b8a6-426b-99c9-3c6488dd27c6] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopActionWithBadCommand/626a6841-b8a6-426b-99c9-3c6488dd27c6] FAILED: org.apache.oozie.action.hadoop.TestSqoopActionExecutor.testSqoopEval Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopEval/5a0416a0-e434-4b80-b9cd-c40eeefe9340] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopEval/5a0416a0-e434-4b80-b9cd-c40eeefe9340] FAILED: org.apache.oozie.action.hadoop.TestSqoopActionExecutor.testSqoopActionWithRedundantPrefix Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopActionWithRedundantPrefix/f1bf5dbf-61a8-4722-b921-1d5c64a5bae0] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopActionWithRedundantPrefix/f1bf5dbf-61a8-4722-b921-1d5c64a5bae0] FAILED: org.apache.oozie.action.hadoop.TestSqoopActionExecutor.testSqoopAction Error Message: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopAction/10ff4ddd-ff25-42e4-8dfc-4ee6939dd5c1] Stack Trace: java.lang.RuntimeException: Could not create testcase dir[/home/jenkins/jenkins-slave/workspace/PreCommit-OOZIE-Build@2/sharelib/sqoop/target/test-data/oozietests/org.apache.oozie.action.hadoop.TestSqoopActionExecutor/testSqoopAction/10ff4ddd-ff25-42e4-8dfc-4ee6939dd5c1]