See <https://builds.apache.org/job/Geode-spark-connector/78/changes>
Changes: [hkhamesra] GEODE-37 In spark connector we call TcpClient static method to get the [klund] GEODE-1906: fix misspelling of Successfully [upthewaterspout] GEODE-1915: Prevent deadlock registering instantiators with gateways ------------------------------------------ [...truncated 1883 lines...] 16/09/23 16:11:05 INFO HttpFileServer: HTTP File server directory is /tmp/spark-f13dac55-087f-4379-aeed-616fbdc7ffac/httpd-02c1fab9-faa0-47f4-b0f3-fd44383eeeb3 16/09/23 16:11:05 INFO HttpServer: Starting HTTP Server 16/09/23 16:11:05 INFO Utils: Successfully started service 'HTTP file server' on port 40135. 16/09/23 16:11:05 INFO SparkEnv: Registering OutputCommitCoordinator 16/09/23 16:11:10 WARN Utils: Service 'SparkUI' could not bind on port 4040. Attempting port 4041. 16/09/23 16:11:15 INFO Utils: Successfully started service 'SparkUI' on port 4041. 16/09/23 16:11:15 INFO SparkUI: Started SparkUI at http://localhost:4041 16/09/23 16:11:15 INFO Executor: Starting executor ID <driver> on host localhost 16/09/23 16:11:15 INFO AkkaUtils: Connecting to HeartbeatReceiver: akka.tcp://sparkDriver@localhost:54872/user/HeartbeatReceiver 16/09/23 16:11:16 INFO NettyBlockTransferService: Server created on 41182 16/09/23 16:11:16 INFO BlockManagerMaster: Trying to register BlockManager 16/09/23 16:11:16 INFO BlockManagerMasterActor: Registering block manager localhost:41182 with 2.8 GB RAM, BlockManagerId(<driver>, localhost, 41182) 16/09/23 16:11:16 INFO BlockManagerMaster: Registered BlockManager === GeodeRunner: stop server 1. === GeodeRunner: stop server 2. [0m[[0minfo[0m] [0m[32mRetrieveRegionIntegrationTest:[0m[0m ...... === GeodeRunner: stop locator ... Successfully stop Geode locator at port 27662. === GeodeRunner: starting locator on port 23825 === GeodeRunner: waiting for locator on port 23825 ....=== GeodeRunner: done waiting for locator on port 23825 === GeodeRunner: starting server1 with clientPort 28993 === GeodeRunner: starting server2 with clientPort 26318 === GeodeRunner: starting server3 with clientPort 29777 === GeodeRunner: starting server4 with clientPort 22946 .... ............................................Locator in /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/locator on hemera.apache.org[23825] as locator is currently online. Process ID: 1860 Uptime: 4 seconds GemFire Version: 1.0.0-incubating-SNAPSHOT Java Version: 1.8.0_66 Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/locator/locator.log JVM Arguments: -Dgemfire.enable-cluster-configuration=true -Dgemfire.load-cluster-configuration-from-dir=false -Dgemfire.jmx-manager-http-port=29684 -Dgemfire.launcher.registerSignalHandlers=true -Djava.awt.headless=true -Dsun.rmi.dgc.server.gcInterval=9223372036854775806 Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-dependencies.jar Successfully connected to: JMX Manager [host=hemera.apache.org, port=1099] Cluster configuration service is up and running. ................ Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server4 on hemera.apache.org[22946] as server4 is currently online. Process ID: 2204 Uptime: 8 seconds GemFire Version: 1.0.0-incubating-SNAPSHOT Java Version: 1.8.0_66 Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server4/server4.log JVM Arguments: -Dgemfire.locators=localhost[23825] -Dgemfire.use-cluster-configuration=true -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSignalHandlers=true -Djava.awt.headless=true -Dsun.rmi.dgc.server.gcInterval=9223372036854775806 Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-dependencies.jar .. Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1 on hemera.apache.org[28993] as server1 is currently online. Process ID: 2199 Uptime: 8 seconds GemFire Version: 1.0.0-incubating-SNAPSHOT Java Version: 1.8.0_66 Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1/server1.log JVM Arguments: -Dgemfire.locators=localhost[23825] -Dgemfire.use-cluster-configuration=true -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSignalHandlers=true -Djava.awt.headless=true -Dsun.rmi.dgc.server.gcInterval=9223372036854775806 Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-dependencies.jar Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2 on hemera.apache.org[26318] as server2 is currently online. Process ID: 2153 Uptime: 9 seconds GemFire Version: 1.0.0-incubating-SNAPSHOT Java Version: 1.8.0_66 Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2/server2.log JVM Arguments: -Dgemfire.locators=localhost[23825] -Dgemfire.use-cluster-configuration=true -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSignalHandlers=true -Djava.awt.headless=true -Dsun.rmi.dgc.server.gcInterval=9223372036854775806 Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-dependencies.jar Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server3 on hemera.apache.org[29777] as server3 is currently online. Process ID: 2175 Uptime: 9 seconds GemFire Version: 1.0.0-incubating-SNAPSHOT Java Version: 1.8.0_66 Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server3/server3.log JVM Arguments: -Dgemfire.locators=localhost[23825] -Dgemfire.use-cluster-configuration=true -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSignalHandlers=true -Djava.awt.headless=true -Dsun.rmi.dgc.server.gcInterval=9223372036854775806 Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-dependencies.jar All WrappedArray(28993, 26318, 29777, 22946).length servers have been started Deploying:geode-functions_2.10-0.5.0.jar 16/09/23 16:11:43 WARN SparkContext: Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext may be running in this JVM (see SPARK-2243). The other SparkContext was created at: org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:61) ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest.setUpBeforeClass(JavaApiIntegrationTest.java:75) sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.lang.reflect.Method.invoke(Method.java:497) org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47) org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44) org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) org.junit.runners.ParentRunner.run(ParentRunner.java:309) org.junit.runners.Suite.runChild(Suite.java:127) org.junit.runners.Suite.runChild(Suite.java:26) org.junit.runners.ParentRunner$3.run(ParentRunner.java:238) org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63) org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236) org.junit.runners.ParentRunner.access$000(ParentRunner.java:53) org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229) org.junit.runners.ParentRunner.run(ParentRunner.java:309) [0m[[0minfo[0m] [0m[31mException encountered when attempting to run a suite with class name: ittest.org.apache.geode.spark.connector.RetrieveRegionIntegrationTest *** ABORTED ***[0m[0m [0m[[0minfo[0m] [0m[31m org.apache.spark.SparkException: Only one SparkContext may be running in this JVM (see SPARK-2243). To ignore this error, set spark.driver.allowMultipleContexts = true. The currently running SparkContext was created at:[0m[0m [0m[[0minfo[0m] [0m[31morg.apache.spark.SparkContext.<init>(SparkContext.scala:80)[0m[0m [0m[[0minfo[0m] [0m[31mittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:50)[0m[0m [0m[[0minfo[0m] [0m[31morg.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)[0m[0m [0m[[0minfo[0m] [0m[31mittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:30)[0m[0m [0m[[0minfo[0m] [0m[31morg.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)[0m[0m [0m[[0minfo[0m] [0m[31mittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30)[0m[0m [0m[[0minfo[0m] [0m[31morg.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:462)[0m[0m [0m[[0minfo[0m] [0m[31morg.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671)[0m[0m [0m[[0minfo[0m] [0m[31msbt.ForkMain$Run$2.call(ForkMain.java:294)[0m[0m [0m[[0minfo[0m] [0m[31msbt.ForkMain$Run$2.call(ForkMain.java:284)[0m[0m [0m[[0minfo[0m] [0m[31mjava.util.concurrent.FutureTask.run(FutureTask.java:266)[0m[0m [0m[[0minfo[0m] [0m[31mjava.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)[0m[0m [0m[[0minfo[0m] [0m[31mjava.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)[0m[0m [0m[[0minfo[0m] [0m[31mjava.lang.Thread.run(Thread.java:745)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1811)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1807)[0m[0m [0m[[0minfo[0m] [0m[31m at scala.Option.foreach(Option.scala:236)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794)[0m[0m [0m[[0minfo[0m] [0m[31m at scala.Option.foreach(Option.scala:236)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$.assertNoOtherContextIsRunning(SparkContext.scala:1794)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$.markPartiallyConstructed(SparkContext.scala:1833)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext.<init>(SparkContext.scala:89)[0m[0m [0m[[0minfo[0m] [0m[31m at ittest.org.apache.geode.spark.connector.RetrieveRegionIntegrationTest.beforeAll(RetrieveRegionIntegrationTest.scala:51)[0m[0m [0m[[0minfo[0m] [0m[31m ...[0m[0m [0m[[0minfo[0m] [0m[32mBasicIntegrationTest:[0m[0m === GeodeRunner: stop server 1. === GeodeRunner: stop server 2. === GeodeRunner: stop server 3. === GeodeRunner: stop server 4. ............ === GeodeRunner: stop locator ... Successfully stop Geode locator at port 23825. === GeodeRunner: starting locator on port 23573 === GeodeRunner: waiting for locator on port 23573 ....=== GeodeRunner: done waiting for locator on port 23573 === GeodeRunner: starting server1 with clientPort 27897 === GeodeRunner: starting server2 with clientPort 20289 .... ....................Locator in /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/locator on hemera.apache.org[23573] as locator is currently online. Process ID: 3273 Uptime: 4 seconds GemFire Version: 1.0.0-incubating-SNAPSHOT Java Version: 1.8.0_66 Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/locator/locator.log JVM Arguments: -Dgemfire.enable-cluster-configuration=true -Dgemfire.load-cluster-configuration-from-dir=false -Dgemfire.jmx-manager-http-port=23053 -Dgemfire.launcher.registerSignalHandlers=true -Djava.awt.headless=true -Dsun.rmi.dgc.server.gcInterval=9223372036854775806 Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-dependencies.jar Successfully connected to: JMX Manager [host=hemera.apache.org, port=1099] Cluster configuration service is up and running. ........ Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2 on hemera.apache.org[20289] as server2 is currently online. Process ID: 3465 Uptime: 7 seconds GemFire Version: 1.0.0-incubating-SNAPSHOT Java Version: 1.8.0_66 Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2/server2.log JVM Arguments: -Dgemfire.locators=localhost[23573] -Dgemfire.use-cluster-configuration=true -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/src/it/resources/test-regions.xml -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSignalHandlers=true -Djava.awt.headless=true -Dsun.rmi.dgc.server.gcInterval=9223372036854775806 Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-dependencies.jar Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1 on hemera.apache.org[27897] as server1 is currently online. Process ID: 3505 Uptime: 7 seconds GemFire Version: 1.0.0-incubating-SNAPSHOT Java Version: 1.8.0_66 Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1/server1.log JVM Arguments: -Dgemfire.locators=localhost[23573] -Dgemfire.use-cluster-configuration=true -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/src/it/resources/test-regions.xml -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSignalHandlers=true -Djava.awt.headless=true -Dsun.rmi.dgc.server.gcInterval=9223372036854775806 Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-assembly/build/install/apache-geode/lib/geode-dependencies.jar All WrappedArray(27897, 20289).length servers have been started Deploying:geode-functions_2.10-0.5.0.jar 16/09/23 16:12:09 WARN SparkContext: Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext may be running in this JVM (see SPARK-2243). The other SparkContext was created at: org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:61) ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest.setUpBeforeClass(JavaApiIntegrationTest.java:75) sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.lang.reflect.Method.invoke(Method.java:497) org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47) org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44) org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) org.junit.runners.ParentRunner.run(ParentRunner.java:309) org.junit.runners.Suite.runChild(Suite.java:127) org.junit.runners.Suite.runChild(Suite.java:26) org.junit.runners.ParentRunner$3.run(ParentRunner.java:238) org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63) org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236) org.junit.runners.ParentRunner.access$000(ParentRunner.java:53) org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229) org.junit.runners.ParentRunner.run(ParentRunner.java:309) [0m[[0minfo[0m] [0m[31mException encountered when attempting to run a suite with class name: ittest.org.apache.geode.spark.connector.BasicIntegrationTest *** ABORTED ***[0m[0m [0m[[0minfo[0m] [0m[31m org.apache.spark.SparkException: Only one SparkContext may be running in this JVM (see SPARK-2243). To ignore this error, set spark.driver.allowMultipleContexts = true. The currently running SparkContext was created at:[0m[0m [0m[[0minfo[0m] [0m[31morg.apache.spark.SparkContext.<init>(SparkContext.scala:80)[0m[0m [0m[[0minfo[0m] [0m[31mittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:50)[0m[0m [0m[[0minfo[0m] [0m[31morg.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)[0m[0m [0m[[0minfo[0m] [0m[31mittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:30)[0m[0m [0m[[0minfo[0m] [0m[31morg.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)[0m[0m [0m[[0minfo[0m] [0m[31mittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30)[0m[0m [0m[[0minfo[0m] [0m[31morg.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:462)[0m[0m [0m[[0minfo[0m] [0m[31morg.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671)[0m[0m [0m[[0minfo[0m] [0m[31msbt.ForkMain$Run$2.call(ForkMain.java:294)[0m[0m [0m[[0minfo[0m] [0m[31msbt.ForkMain$Run$2.call(ForkMain.java:284)[0m[0m [0m[[0minfo[0m] [0m[31mjava.util.concurrent.FutureTask.run(FutureTask.java:266)[0m[0m [0m[[0minfo[0m] [0m[31mjava.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)[0m[0m [0m[[0minfo[0m] [0m[31mjava.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)[0m[0m [0m[[0minfo[0m] [0m[31mjava.lang.Thread.run(Thread.java:745)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1811)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1807)[0m[0m [0m[[0minfo[0m] [0m[31m at scala.Option.foreach(Option.scala:236)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794)[0m[0m [0m[[0minfo[0m] [0m[31m at scala.Option.foreach(Option.scala:236)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$.assertNoOtherContextIsRunning(SparkContext.scala:1794)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext$.markPartiallyConstructed(SparkContext.scala:1833)[0m[0m [0m[[0minfo[0m] [0m[31m at org.apache.spark.SparkContext.<init>(SparkContext.scala:89)[0m[0m [0m[[0minfo[0m] [0m[31m at ittest.org.apache.geode.spark.connector.BasicIntegrationTest.beforeAll(BasicIntegrationTest.scala:58)[0m[0m [0m[[0minfo[0m] [0m[31m ...[0m[0m [0m[[0minfo[0m] [0mScalaTest[0m [0m[[0minfo[0m] [0m[36mRun completed in 1 minute, 59 seconds.[0m[0m [0m[[0minfo[0m] [0m[36mTotal number of tests run: 0[0m[0m [0m[[0minfo[0m] [0m[36mSuites: completed 1, aborted 3[0m[0m [0m[[0minfo[0m] [0m[36mTests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0[0m[0m [0m[[0minfo[0m] [0m[31m*** 3 SUITES ABORTED ***[0m[0m [0m[[31merror[0m] [0mError: Total 3, Failed 0, Errors 3, Passed 0[0m [0m[[31merror[0m] [0mError during tests:[0m [0m[[31merror[0m] [0m ittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest[0m [0m[[31merror[0m] [0m ittest.org.apache.geode.spark.connector.RetrieveRegionIntegrationTest[0m [0m[[31merror[0m] [0m ittest.org.apache.geode.spark.connector.BasicIntegrationTest[0m [0m[[31merror[0m] [0m(geode-spark-connector/it:[31mtest[0m) sbt.TestsFailedException: Tests unsuccessful[0m [0m[[31merror[0m] [0mTotal time: 128 s, completed Sep 23, 2016 4:12:09 PM[0m Build step 'Execute shell' marked build as failure Recording test results Skipped archiving because build is not successful