I created GEODE-1934 for this. It looks like the problem is actually that
our dependencies for geode-core are messed up. spring-core is marked
optional, but we're using it in critical places like this
SSLConfigurationFactory.

In my opinion we shouldn't depend on spring-core at all unless we're
actually going to use it for things other than StringUtils. I think we've
accidentally introduced dependencies on it because the gfsh code in the
core is pulling in a bunch of spring libraries.

-Dan


On Fri, Sep 23, 2016 at 9:12 AM, Apache Jenkins Server <
jenk...@builds.apache.org> wrote:

> See <https://builds.apache.org/job/Geode-spark-connector/78/changes>
>
> Changes:
>
> [hkhamesra] GEODE-37 In spark connector we call TcpClient static method to
> get the
>
> [klund] GEODE-1906: fix misspelling of Successfully
>
> [upthewaterspout] GEODE-1915: Prevent deadlock registering instantiators
> with gateways
>
> ------------------------------------------
> [...truncated 1883 lines...]
> 16/09/23 16:11:05 INFO HttpFileServer: HTTP File server directory is
> /tmp/spark-f13dac55-087f-4379-aeed-616fbdc7ffac/httpd-
> 02c1fab9-faa0-47f4-b0f3-fd44383eeeb3
> 16/09/23 16:11:05 INFO HttpServer: Starting HTTP Server
> 16/09/23 16:11:05 INFO Utils: Successfully started service 'HTTP file
> server' on port 40135.
> 16/09/23 16:11:05 INFO SparkEnv: Registering OutputCommitCoordinator
> 16/09/23 16:11:10 WARN Utils: Service 'SparkUI' could not bind on port
> 4040. Attempting port 4041.
> 16/09/23 16:11:15 INFO Utils: Successfully started service 'SparkUI' on
> port 4041.
> 16/09/23 16:11:15 INFO SparkUI: Started SparkUI at http://localhost:4041
> 16/09/23 16:11:15 INFO Executor: Starting executor ID <driver> on host
> localhost
> 16/09/23 16:11:15 INFO AkkaUtils: Connecting to HeartbeatReceiver:
> akka.tcp://sparkDriver@localhost:54872/user/HeartbeatReceiver
> 16/09/23 16:11:16 INFO NettyBlockTransferService: Server created on 41182
> 16/09/23 16:11:16 INFO BlockManagerMaster: Trying to register BlockManager
> 16/09/23 16:11:16 INFO BlockManagerMasterActor: Registering block manager
> localhost:41182 with 2.8 GB RAM, BlockManagerId(<driver>, localhost, 41182)
> 16/09/23 16:11:16 INFO BlockManagerMaster: Registered BlockManager
> === GeodeRunner: stop server 1.
> === GeodeRunner: stop server 2.
>  [0m[ [0minfo [0m]  [0m [32mRetrieveRegionIntegrationTest: [0m [0m
> ......
>
> === GeodeRunner: stop locator
> ...
> Successfully stop Geode locator at port 27662.
> === GeodeRunner: starting locator on port 23825
> === GeodeRunner: waiting for locator on port 23825
> ....=== GeodeRunner: done waiting for locator on port 23825
> === GeodeRunner: starting server1 with clientPort 28993
> === GeodeRunner: starting server2 with clientPort 26318
> === GeodeRunner: starting server3 with clientPort 29777
> === GeodeRunner: starting server4 with clientPort 22946
> ....
> ............................................Locator in
> /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
> connector/geode-spark-connector/target/testgeode/locator on
> hemera.apache.org[23825] as locator is currently online.
> Process ID: 1860
> Uptime: 4 seconds
> GemFire Version: 1.0.0-incubating-SNAPSHOT
> Java Version: 1.8.0_66
> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
> locator/locator.log
> JVM Arguments: -Dgemfire.enable-cluster-configuration=true
> -Dgemfire.load-cluster-configuration-from-dir=false
> -Dgemfire.jmx-manager-http-port=29684 
> -Dgemfire.launcher.registerSignalHandlers=true
> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
> gcInterval=9223372036854775806
> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/
> lib/geode-dependencies.jar
>
> Successfully connected to: JMX Manager [host=hemera.apache.org, port=1099]
>
> Cluster configuration service is up and running.
>
> ................
> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server4
> on hemera.apache.org[22946] as server4 is currently online.
> Process ID: 2204
> Uptime: 8 seconds
> GemFire Version: 1.0.0-incubating-SNAPSHOT
> Java Version: 1.8.0_66
> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
> server4/server4.log
> JVM Arguments: -Dgemfire.locators=localhost[23825] 
> -Dgemfire.use-cluster-configuration=true
> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml
> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
> -XX:OnOutOfMemoryError=kill -KILL %p 
> -Dgemfire.launcher.registerSignalHandlers=true
> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
> gcInterval=9223372036854775806
> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
> spark-connector/geode-assembly/build/install/apache-
> geode/lib/geode-dependencies.jar
>
> ..
> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1
> on hemera.apache.org[28993] as server1 is currently online.
> Process ID: 2199
> Uptime: 8 seconds
> GemFire Version: 1.0.0-incubating-SNAPSHOT
> Java Version: 1.8.0_66
> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
> server1/server1.log
> JVM Arguments: -Dgemfire.locators=localhost[23825] 
> -Dgemfire.use-cluster-configuration=true
> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml
> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
> -XX:OnOutOfMemoryError=kill -KILL %p 
> -Dgemfire.launcher.registerSignalHandlers=true
> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
> gcInterval=9223372036854775806
> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
> spark-connector/geode-assembly/build/install/apache-
> geode/lib/geode-dependencies.jar
>
>
>
> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2
> on hemera.apache.org[26318] as server2 is currently online.
> Process ID: 2153
> Uptime: 9 seconds
> GemFire Version: 1.0.0-incubating-SNAPSHOT
> Java Version: 1.8.0_66
> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
> server2/server2.log
> JVM Arguments: -Dgemfire.locators=localhost[23825] 
> -Dgemfire.use-cluster-configuration=true
> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml
> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
> -XX:OnOutOfMemoryError=kill -KILL %p 
> -Dgemfire.launcher.registerSignalHandlers=true
> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
> gcInterval=9223372036854775806
> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
> spark-connector/geode-assembly/build/install/apache-
> geode/lib/geode-dependencies.jar
>
> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server3
> on hemera.apache.org[29777] as server3 is currently online.
> Process ID: 2175
> Uptime: 9 seconds
> GemFire Version: 1.0.0-incubating-SNAPSHOT
> Java Version: 1.8.0_66
> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
> server3/server3.log
> JVM Arguments: -Dgemfire.locators=localhost[23825] 
> -Dgemfire.use-cluster-configuration=true
> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml
> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
> -XX:OnOutOfMemoryError=kill -KILL %p 
> -Dgemfire.launcher.registerSignalHandlers=true
> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
> gcInterval=9223372036854775806
> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
> spark-connector/geode-assembly/build/install/apache-
> geode/lib/geode-dependencies.jar
>
> All WrappedArray(28993, 26318, 29777, 22946).length servers have been
> started
> Deploying:geode-functions_2.10-0.5.0.jar
> 16/09/23 16:11:43 WARN SparkContext: Another SparkContext is being
> constructed (or threw an exception in its constructor).  This may indicate
> an error, since only one SparkContext may be running in this JVM (see
> SPARK-2243). The other SparkContext was created at:
> org.apache.spark.api.java.JavaSparkContext.<init>(
> JavaSparkContext.scala:61)
> ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest.
> setUpBeforeClass(JavaApiIntegrationTest.java:75)
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:
> 62)
> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
> java.lang.reflect.Method.invoke(Method.java:497)
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(
> FrameworkMethod.java:47)
> org.junit.internal.runners.model.ReflectiveCallable.run(
> ReflectiveCallable.java:12)
> org.junit.runners.model.FrameworkMethod.invokeExplosively(
> FrameworkMethod.java:44)
> org.junit.internal.runners.statements.RunBefores.
> evaluate(RunBefores.java:24)
> org.junit.internal.runners.statements.RunAfters.evaluate(
> RunAfters.java:27)
> org.junit.runners.ParentRunner.run(ParentRunner.java:309)
> org.junit.runners.Suite.runChild(Suite.java:127)
> org.junit.runners.Suite.runChild(Suite.java:26)
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
> org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>  [0m[ [0minfo [0m]  [0m [31mException encountered when attempting to run a
> suite with class name: 
> ittest.org.apache.geode.spark.connector.RetrieveRegionIntegrationTest
> *** ABORTED *** [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  org.apache.spark.SparkException: Only one
> SparkContext may be running in this JVM (see SPARK-2243). To ignore this
> error, set spark.driver.allowMultipleContexts = true. The currently
> running SparkContext was created at: [0m [0m
>  [0m[ [0minfo [0m]  [0m 
> [31morg.apache.spark.SparkContext.<init>(SparkContext.scala:80)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:50)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31morg.scalatest.BeforeAndAfterAll$class.
> beforeAll(BeforeAndAfterAll.scala:187) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:30)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m 
> [31morg.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
> RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31morg.scalatest.tools.Framework.org
> $scalatest$tools$Framework$$runSuite(Framework.scala:462) [0m [0m
>  [0m[ [0minfo [0m]  [0m 
> [31morg.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:294)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:284)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
> FutureTask.run(FutureTask.java:266) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
> ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
> ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mjava.lang.Thread.run(Thread.java:745) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1811)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1807)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at scala.Option.foreach(Option.scala:236)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at scala.Option.foreach(Option.scala:236)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$.
> assertNoOtherContextIsRunning(SparkContext.scala:1794) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$
> .markPartiallyConstructed(SparkContext.scala:1833) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at 
> org.apache.spark.SparkContext.<init>(SparkContext.scala:89)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at ittest.org.apache.geode.spark.connector.
> RetrieveRegionIntegrationTest.beforeAll(RetrieveRegionIntegrationTest.scala:51)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  ... [0m [0m
>  [0m[ [0minfo [0m]  [0m [32mBasicIntegrationTest: [0m [0m
> === GeodeRunner: stop server 1.
> === GeodeRunner: stop server 2.
> === GeodeRunner: stop server 3.
> === GeodeRunner: stop server 4.
> ............
>
>
>
> === GeodeRunner: stop locator
> ...
> Successfully stop Geode locator at port 23825.
> === GeodeRunner: starting locator on port 23573
> === GeodeRunner: waiting for locator on port 23573
> ....=== GeodeRunner: done waiting for locator on port 23573
> === GeodeRunner: starting server1 with clientPort 27897
> === GeodeRunner: starting server2 with clientPort 20289
> ....
> ....................Locator in /x1/jenkins/jenkins-slave/
> workspace/Geode-spark-connector/geode-spark-connector/geode-spark-
> connector/target/testgeode/locator on hemera.apache.org[23573] as locator
> is currently online.
> Process ID: 3273
> Uptime: 4 seconds
> GemFire Version: 1.0.0-incubating-SNAPSHOT
> Java Version: 1.8.0_66
> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
> locator/locator.log
> JVM Arguments: -Dgemfire.enable-cluster-configuration=true
> -Dgemfire.load-cluster-configuration-from-dir=false
> -Dgemfire.jmx-manager-http-port=23053 
> -Dgemfire.launcher.registerSignalHandlers=true
> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
> gcInterval=9223372036854775806
> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/
> lib/geode-dependencies.jar
>
> Successfully connected to: JMX Manager [host=hemera.apache.org, port=1099]
>
> Cluster configuration service is up and running.
>
> ........
> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2
> on hemera.apache.org[20289] as server2 is currently online.
> Process ID: 3465
> Uptime: 7 seconds
> GemFire Version: 1.0.0-incubating-SNAPSHOT
> Java Version: 1.8.0_66
> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
> server2/server2.log
> JVM Arguments: -Dgemfire.locators=localhost[23573] 
> -Dgemfire.use-cluster-configuration=true
> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
> connector/geode-spark-connector/src/it/resources/test-regions.xml
> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
> -XX:OnOutOfMemoryError=kill -KILL %p 
> -Dgemfire.launcher.registerSignalHandlers=true
> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
> gcInterval=9223372036854775806
> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
> spark-connector/geode-assembly/build/install/apache-
> geode/lib/geode-dependencies.jar
>
>
> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1
> on hemera.apache.org[27897] as server1 is currently online.
> Process ID: 3505
> Uptime: 7 seconds
> GemFire Version: 1.0.0-incubating-SNAPSHOT
> Java Version: 1.8.0_66
> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
> server1/server1.log
> JVM Arguments: -Dgemfire.locators=localhost[23573] 
> -Dgemfire.use-cluster-configuration=true
> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
> connector/geode-spark-connector/src/it/resources/test-regions.xml
> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
> -XX:OnOutOfMemoryError=kill -KILL %p 
> -Dgemfire.launcher.registerSignalHandlers=true
> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
> gcInterval=9223372036854775806
> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
> spark-connector/geode-assembly/build/install/apache-
> geode/lib/geode-dependencies.jar
>
> All WrappedArray(27897, 20289).length servers have been started
> Deploying:geode-functions_2.10-0.5.0.jar
> 16/09/23 16:12:09 WARN SparkContext: Another SparkContext is being
> constructed (or threw an exception in its constructor).  This may indicate
> an error, since only one SparkContext may be running in this JVM (see
> SPARK-2243). The other SparkContext was created at:
> org.apache.spark.api.java.JavaSparkContext.<init>(
> JavaSparkContext.scala:61)
> ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest.
> setUpBeforeClass(JavaApiIntegrationTest.java:75)
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:
> 62)
> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
> java.lang.reflect.Method.invoke(Method.java:497)
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(
> FrameworkMethod.java:47)
> org.junit.internal.runners.model.ReflectiveCallable.run(
> ReflectiveCallable.java:12)
> org.junit.runners.model.FrameworkMethod.invokeExplosively(
> FrameworkMethod.java:44)
> org.junit.internal.runners.statements.RunBefores.
> evaluate(RunBefores.java:24)
> org.junit.internal.runners.statements.RunAfters.evaluate(
> RunAfters.java:27)
> org.junit.runners.ParentRunner.run(ParentRunner.java:309)
> org.junit.runners.Suite.runChild(Suite.java:127)
> org.junit.runners.Suite.runChild(Suite.java:26)
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
> org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>  [0m[ [0minfo [0m]  [0m [31mException encountered when attempting to run a
> suite with class name: 
> ittest.org.apache.geode.spark.connector.BasicIntegrationTest
> *** ABORTED *** [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  org.apache.spark.SparkException: Only one
> SparkContext may be running in this JVM (see SPARK-2243). To ignore this
> error, set spark.driver.allowMultipleContexts = true. The currently
> running SparkContext was created at: [0m [0m
>  [0m[ [0minfo [0m]  [0m 
> [31morg.apache.spark.SparkContext.<init>(SparkContext.scala:80)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:50)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31morg.scalatest.BeforeAndAfterAll$class.
> beforeAll(BeforeAndAfterAll.scala:187) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:30)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m 
> [31morg.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
> RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31morg.scalatest.tools.Framework.org
> $scalatest$tools$Framework$$runSuite(Framework.scala:462) [0m [0m
>  [0m[ [0minfo [0m]  [0m 
> [31morg.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:294)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:284)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
> FutureTask.run(FutureTask.java:266) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
> ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
> ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31mjava.lang.Thread.run(Thread.java:745) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1811)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1807)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at scala.Option.foreach(Option.scala:236)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at scala.Option.foreach(Option.scala:236)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$.
> assertNoOtherContextIsRunning(SparkContext.scala:1794) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$
> .markPartiallyConstructed(SparkContext.scala:1833) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at 
> org.apache.spark.SparkContext.<init>(SparkContext.scala:89)
> [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  at ittest.org.apache.geode.spark.connector.
> BasicIntegrationTest.beforeAll(BasicIntegrationTest.scala:58) [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m  ... [0m [0m
>  [0m[ [0minfo [0m]  [0mScalaTest [0m
>  [0m[ [0minfo [0m]  [0m [36mRun completed in 1 minute, 59 seconds. [0m [0m
>  [0m[ [0minfo [0m]  [0m [36mTotal number of tests run: 0 [0m [0m
>  [0m[ [0minfo [0m]  [0m [36mSuites: completed 1, aborted 3 [0m [0m
>  [0m[ [0minfo [0m]  [0m [36mTests: succeeded 0, failed 0, canceled 0,
> ignored 0, pending 0 [0m [0m
>  [0m[ [0minfo [0m]  [0m [31m*** 3 SUITES ABORTED *** [0m [0m
>  [0m[ [31merror [0m]  [0mError: Total 3, Failed 0, Errors 3, Passed 0 [0m
>  [0m[ [31merror [0m]  [0mError during tests: [0m
>  [0m[ [31merror [0m]  [0m       
> ittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest
> [0m
>  [0m[ [31merror [0m]  [0m       
> ittest.org.apache.geode.spark.connector.RetrieveRegionIntegrationTest
> [0m
>  [0m[ [31merror [0m]  [0m       
> ittest.org.apache.geode.spark.connector.BasicIntegrationTest
> [0m
>  [0m[ [31merror [0m]  [0m(geode-spark-connector/it: [31mtest [0m)
> sbt.TestsFailedException: Tests unsuccessful [0m
>  [0m[ [31merror [0m]  [0mTotal time: 128 s, completed Sep 23, 2016 4:12:09
> PM [0m
> Build step 'Execute shell' marked build as failure
> Recording test results
> Skipped archiving because build is not successful
>

Reply via email to