Yep, I’m seeing failures on any client app that doesn’t explicitly include 
spring as dependency.

Exception in thread "main" java.lang.NoClassDefFoundError: 
org/springframework/util/StringUtils
        at 
org.apache.geode.internal.net.SSLConfigurationFactory.configureSSLPropertiesFromSystemProperties(SSLConfigurationFactory.java:274)
        at 
org.apache.geode.internal.net.SSLConfigurationFactory.configureSSLPropertiesFromSystemProperties(SSLConfigurationFactory.java:270)
        at 
org.apache.geode.internal.net.SSLConfigurationFactory.createSSLConfigForComponent(SSLConfigurationFactory.java:138)
        at 
org.apache.geode.internal.net.SSLConfigurationFactory.getSSLConfigForComponent(SSLConfigurationFactory.java:67)
        at 
org.apache.geode.internal.net.SocketCreatorFactory.getSocketCreatorForComponent(SocketCreatorFactory.java:67)
        at 
org.apache.geode.distributed.internal.tcpserver.TcpClient.<init>(TcpClient.java:69)
        at 
org.apache.geode.cache.client.internal.AutoConnectionSourceImpl.<init>(AutoConnectionSourceImpl.java:114)
        at 
org.apache.geode.cache.client.internal.PoolImpl.getSourceImpl(PoolImpl.java:579)
        at 
org.apache.geode.cache.client.internal.PoolImpl.<init>(PoolImpl.java:219)
        at 
org.apache.geode.cache.client.internal.PoolImpl.create(PoolImpl.java:132)
        at 
org.apache.geode.internal.cache.PoolFactoryImpl.create(PoolFactoryImpl.java:319)
        at 
org.apache.geode.internal.cache.GemFireCacheImpl.determineDefaultPool(GemFireCacheImpl.java:2943)
        at 
org.apache.geode.internal.cache.GemFireCacheImpl.initializeDeclarativeCache(GemFireCacheImpl.java:1293)
        at 
org.apache.geode.internal.cache.GemFireCacheImpl.initialize(GemFireCacheImpl.java:1124)
        at 
org.apache.geode.internal.cache.GemFireCacheImpl.basicCreate(GemFireCacheImpl.java:765)
        at 
org.apache.geode.internal.cache.GemFireCacheImpl.createClient(GemFireCacheImpl.java:740)
        at 
org.apache.geode.cache.client.ClientCacheFactory.basicCreate(ClientCacheFactory.java:235)
        at 
org.apache.geode.cache.client.ClientCacheFactory.create(ClientCacheFactory.java:189)
        at HelloWorld.main(HelloWorld.java:25)
Caused by: java.lang.ClassNotFoundException: 
org.springframework.util.StringUtils
        at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
        ... 19 more

Anthony


> On Sep 23, 2016, at 4:34 PM, Dan Smith <dsm...@pivotal.io> wrote:
> 
> I created GEODE-1934 for this. It looks like the problem is actually that
> our dependencies for geode-core are messed up. spring-core is marked
> optional, but we're using it in critical places like this
> SSLConfigurationFactory.
> 
> In my opinion we shouldn't depend on spring-core at all unless we're
> actually going to use it for things other than StringUtils. I think we've
> accidentally introduced dependencies on it because the gfsh code in the
> core is pulling in a bunch of spring libraries.
> 
> -Dan
> 
> 
> On Fri, Sep 23, 2016 at 9:12 AM, Apache Jenkins Server <
> jenk...@builds.apache.org> wrote:
> 
>> See <https://builds.apache.org/job/Geode-spark-connector/78/changes>
>> 
>> Changes:
>> 
>> [hkhamesra] GEODE-37 In spark connector we call TcpClient static method to
>> get the
>> 
>> [klund] GEODE-1906: fix misspelling of Successfully
>> 
>> [upthewaterspout] GEODE-1915: Prevent deadlock registering instantiators
>> with gateways
>> 
>> ------------------------------------------
>> [...truncated 1883 lines...]
>> 16/09/23 16:11:05 INFO HttpFileServer: HTTP File server directory is
>> /tmp/spark-f13dac55-087f-4379-aeed-616fbdc7ffac/httpd-
>> 02c1fab9-faa0-47f4-b0f3-fd44383eeeb3
>> 16/09/23 16:11:05 INFO HttpServer: Starting HTTP Server
>> 16/09/23 16:11:05 INFO Utils: Successfully started service 'HTTP file
>> server' on port 40135.
>> 16/09/23 16:11:05 INFO SparkEnv: Registering OutputCommitCoordinator
>> 16/09/23 16:11:10 WARN Utils: Service 'SparkUI' could not bind on port
>> 4040. Attempting port 4041.
>> 16/09/23 16:11:15 INFO Utils: Successfully started service 'SparkUI' on
>> port 4041.
>> 16/09/23 16:11:15 INFO SparkUI: Started SparkUI at http://localhost:4041
>> 16/09/23 16:11:15 INFO Executor: Starting executor ID <driver> on host
>> localhost
>> 16/09/23 16:11:15 INFO AkkaUtils: Connecting to HeartbeatReceiver:
>> akka.tcp://sparkDriver@localhost:54872/user/HeartbeatReceiver
>> 16/09/23 16:11:16 INFO NettyBlockTransferService: Server created on 41182
>> 16/09/23 16:11:16 INFO BlockManagerMaster: Trying to register BlockManager
>> 16/09/23 16:11:16 INFO BlockManagerMasterActor: Registering block manager
>> localhost:41182 with 2.8 GB RAM, BlockManagerId(<driver>, localhost, 41182)
>> 16/09/23 16:11:16 INFO BlockManagerMaster: Registered BlockManager
>> === GeodeRunner: stop server 1.
>> === GeodeRunner: stop server 2.
>> [0m[ [0minfo [0m]  [0m [32mRetrieveRegionIntegrationTest: [0m [0m
>> ......
>> 
>> === GeodeRunner: stop locator
>> ...
>> Successfully stop Geode locator at port 27662.
>> === GeodeRunner: starting locator on port 23825
>> === GeodeRunner: waiting for locator on port 23825
>> ....=== GeodeRunner: done waiting for locator on port 23825
>> === GeodeRunner: starting server1 with clientPort 28993
>> === GeodeRunner: starting server2 with clientPort 26318
>> === GeodeRunner: starting server3 with clientPort 29777
>> === GeodeRunner: starting server4 with clientPort 22946
>> ....
>> ............................................Locator in
>> /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
>> connector/geode-spark-connector/target/testgeode/locator on
>> hemera.apache.org[23825] as locator is currently online.
>> Process ID: 1860
>> Uptime: 4 seconds
>> GemFire Version: 1.0.0-incubating-SNAPSHOT
>> Java Version: 1.8.0_66
>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
>> locator/locator.log
>> JVM Arguments: -Dgemfire.enable-cluster-configuration=true
>> -Dgemfire.load-cluster-configuration-from-dir=false
>> -Dgemfire.jmx-manager-http-port=29684 
>> -Dgemfire.launcher.registerSignalHandlers=true
>> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
>> gcInterval=9223372036854775806
>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/
>> lib/geode-dependencies.jar
>> 
>> Successfully connected to: JMX Manager [host=hemera.apache.org, port=1099]
>> 
>> Cluster configuration service is up and running.
>> 
>> ................
>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server4
>> on hemera.apache.org[22946] as server4 is currently online.
>> Process ID: 2204
>> Uptime: 8 seconds
>> GemFire Version: 1.0.0-incubating-SNAPSHOT
>> Java Version: 1.8.0_66
>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
>> server4/server4.log
>> JVM Arguments: -Dgemfire.locators=localhost[23825] 
>> -Dgemfire.use-cluster-configuration=true
>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
>> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml
>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
>> -XX:OnOutOfMemoryError=kill -KILL %p 
>> -Dgemfire.launcher.registerSignalHandlers=true
>> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
>> gcInterval=9223372036854775806
>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
>> spark-connector/geode-assembly/build/install/apache-
>> geode/lib/geode-dependencies.jar
>> 
>> ..
>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1
>> on hemera.apache.org[28993] as server1 is currently online.
>> Process ID: 2199
>> Uptime: 8 seconds
>> GemFire Version: 1.0.0-incubating-SNAPSHOT
>> Java Version: 1.8.0_66
>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
>> server1/server1.log
>> JVM Arguments: -Dgemfire.locators=localhost[23825] 
>> -Dgemfire.use-cluster-configuration=true
>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
>> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml
>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
>> -XX:OnOutOfMemoryError=kill -KILL %p 
>> -Dgemfire.launcher.registerSignalHandlers=true
>> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
>> gcInterval=9223372036854775806
>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
>> spark-connector/geode-assembly/build/install/apache-
>> geode/lib/geode-dependencies.jar
>> 
>> 
>> 
>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2
>> on hemera.apache.org[26318] as server2 is currently online.
>> Process ID: 2153
>> Uptime: 9 seconds
>> GemFire Version: 1.0.0-incubating-SNAPSHOT
>> Java Version: 1.8.0_66
>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
>> server2/server2.log
>> JVM Arguments: -Dgemfire.locators=localhost[23825] 
>> -Dgemfire.use-cluster-configuration=true
>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
>> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml
>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
>> -XX:OnOutOfMemoryError=kill -KILL %p 
>> -Dgemfire.launcher.registerSignalHandlers=true
>> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
>> gcInterval=9223372036854775806
>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
>> spark-connector/geode-assembly/build/install/apache-
>> geode/lib/geode-dependencies.jar
>> 
>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server3
>> on hemera.apache.org[29777] as server3 is currently online.
>> Process ID: 2175
>> Uptime: 9 seconds
>> GemFire Version: 1.0.0-incubating-SNAPSHOT
>> Java Version: 1.8.0_66
>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
>> server3/server3.log
>> JVM Arguments: -Dgemfire.locators=localhost[23825] 
>> -Dgemfire.use-cluster-configuration=true
>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
>> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml
>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
>> -XX:OnOutOfMemoryError=kill -KILL %p 
>> -Dgemfire.launcher.registerSignalHandlers=true
>> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
>> gcInterval=9223372036854775806
>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
>> spark-connector/geode-assembly/build/install/apache-
>> geode/lib/geode-dependencies.jar
>> 
>> All WrappedArray(28993, 26318, 29777, 22946).length servers have been
>> started
>> Deploying:geode-functions_2.10-0.5.0.jar
>> 16/09/23 16:11:43 WARN SparkContext: Another SparkContext is being
>> constructed (or threw an exception in its constructor).  This may indicate
>> an error, since only one SparkContext may be running in this JVM (see
>> SPARK-2243). The other SparkContext was created at:
>> org.apache.spark.api.java.JavaSparkContext.<init>(
>> JavaSparkContext.scala:61)
>> ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest.
>> setUpBeforeClass(JavaApiIntegrationTest.java:75)
>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:
>> 62)
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
>> DelegatingMethodAccessorImpl.java:43)
>> java.lang.reflect.Method.invoke(Method.java:497)
>> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(
>> FrameworkMethod.java:47)
>> org.junit.internal.runners.model.ReflectiveCallable.run(
>> ReflectiveCallable.java:12)
>> org.junit.runners.model.FrameworkMethod.invokeExplosively(
>> FrameworkMethod.java:44)
>> org.junit.internal.runners.statements.RunBefores.
>> evaluate(RunBefores.java:24)
>> org.junit.internal.runners.statements.RunAfters.evaluate(
>> RunAfters.java:27)
>> org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>> org.junit.runners.Suite.runChild(Suite.java:127)
>> org.junit.runners.Suite.runChild(Suite.java:26)
>> org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>> org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>> org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>> [0m[ [0minfo [0m]  [0m [31mException encountered when attempting to run a
>> suite with class name: 
>> ittest.org.apache.geode.spark.connector.RetrieveRegionIntegrationTest
>> *** ABORTED *** [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  org.apache.spark.SparkException: Only one
>> SparkContext may be running in this JVM (see SPARK-2243). To ignore this
>> error, set spark.driver.allowMultipleContexts = true. The currently
>> running SparkContext was created at: [0m [0m
>> [0m[ [0minfo [0m]  [0m 
>> [31morg.apache.spark.SparkContext.<init>(SparkContext.scala:80)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
>> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:50)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31morg.scalatest.BeforeAndAfterAll$class.
>> beforeAll(BeforeAndAfterAll.scala:187) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
>> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:30)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m 
>> [31morg.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
>> RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31morg.scalatest.tools.Framework.org
>> $scalatest$tools$Framework$$runSuite(Framework.scala:462) [0m [0m
>> [0m[ [0minfo [0m]  [0m 
>> [31morg.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:294)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:284)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
>> FutureTask.run(FutureTask.java:266) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
>> ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
>> ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mjava.lang.Thread.run(Thread.java:745) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
>> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1811)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
>> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1807)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at scala.Option.foreach(Option.scala:236)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
>> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
>> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at scala.Option.foreach(Option.scala:236)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$.
>> assertNoOtherContextIsRunning(SparkContext.scala:1794) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$
>> .markPartiallyConstructed(SparkContext.scala:1833) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at 
>> org.apache.spark.SparkContext.<init>(SparkContext.scala:89)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at ittest.org.apache.geode.spark.connector.
>> RetrieveRegionIntegrationTest.beforeAll(RetrieveRegionIntegrationTest.scala:51)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  ... [0m [0m
>> [0m[ [0minfo [0m]  [0m [32mBasicIntegrationTest: [0m [0m
>> === GeodeRunner: stop server 1.
>> === GeodeRunner: stop server 2.
>> === GeodeRunner: stop server 3.
>> === GeodeRunner: stop server 4.
>> ............
>> 
>> 
>> 
>> === GeodeRunner: stop locator
>> ...
>> Successfully stop Geode locator at port 23825.
>> === GeodeRunner: starting locator on port 23573
>> === GeodeRunner: waiting for locator on port 23573
>> ....=== GeodeRunner: done waiting for locator on port 23573
>> === GeodeRunner: starting server1 with clientPort 27897
>> === GeodeRunner: starting server2 with clientPort 20289
>> ....
>> ....................Locator in /x1/jenkins/jenkins-slave/
>> workspace/Geode-spark-connector/geode-spark-connector/geode-spark-
>> connector/target/testgeode/locator on hemera.apache.org[23573] as locator
>> is currently online.
>> Process ID: 3273
>> Uptime: 4 seconds
>> GemFire Version: 1.0.0-incubating-SNAPSHOT
>> Java Version: 1.8.0_66
>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
>> locator/locator.log
>> JVM Arguments: -Dgemfire.enable-cluster-configuration=true
>> -Dgemfire.load-cluster-configuration-from-dir=false
>> -Dgemfire.jmx-manager-http-port=23053 
>> -Dgemfire.launcher.registerSignalHandlers=true
>> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
>> gcInterval=9223372036854775806
>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/
>> lib/geode-dependencies.jar
>> 
>> Successfully connected to: JMX Manager [host=hemera.apache.org, port=1099]
>> 
>> Cluster configuration service is up and running.
>> 
>> ........
>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2
>> on hemera.apache.org[20289] as server2 is currently online.
>> Process ID: 3465
>> Uptime: 7 seconds
>> GemFire Version: 1.0.0-incubating-SNAPSHOT
>> Java Version: 1.8.0_66
>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
>> server2/server2.log
>> JVM Arguments: -Dgemfire.locators=localhost[23573] 
>> -Dgemfire.use-cluster-configuration=true
>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
>> connector/geode-spark-connector/src/it/resources/test-regions.xml
>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
>> -XX:OnOutOfMemoryError=kill -KILL %p 
>> -Dgemfire.launcher.registerSignalHandlers=true
>> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
>> gcInterval=9223372036854775806
>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
>> spark-connector/geode-assembly/build/install/apache-
>> geode/lib/geode-dependencies.jar
>> 
>> 
>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1
>> on hemera.apache.org[27897] as server1 is currently online.
>> Process ID: 3505
>> Uptime: 7 seconds
>> GemFire Version: 1.0.0-incubating-SNAPSHOT
>> Java Version: 1.8.0_66
>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/
>> server1/server1.log
>> JVM Arguments: -Dgemfire.locators=localhost[23573] 
>> -Dgemfire.use-cluster-configuration=true
>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/
>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark-
>> connector/geode-spark-connector/src/it/resources/test-regions.xml
>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false
>> -XX:OnOutOfMemoryError=kill -KILL %p 
>> -Dgemfire.launcher.registerSignalHandlers=true
>> -Djava.awt.headless=true -Dsun.rmi.dgc.server.
>> gcInterval=9223372036854775806
>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0-
>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark-
>> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/
>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode-
>> spark-connector/geode-assembly/build/install/apache-
>> geode/lib/geode-dependencies.jar
>> 
>> All WrappedArray(27897, 20289).length servers have been started
>> Deploying:geode-functions_2.10-0.5.0.jar
>> 16/09/23 16:12:09 WARN SparkContext: Another SparkContext is being
>> constructed (or threw an exception in its constructor).  This may indicate
>> an error, since only one SparkContext may be running in this JVM (see
>> SPARK-2243). The other SparkContext was created at:
>> org.apache.spark.api.java.JavaSparkContext.<init>(
>> JavaSparkContext.scala:61)
>> ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest.
>> setUpBeforeClass(JavaApiIntegrationTest.java:75)
>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:
>> 62)
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
>> DelegatingMethodAccessorImpl.java:43)
>> java.lang.reflect.Method.invoke(Method.java:497)
>> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(
>> FrameworkMethod.java:47)
>> org.junit.internal.runners.model.ReflectiveCallable.run(
>> ReflectiveCallable.java:12)
>> org.junit.runners.model.FrameworkMethod.invokeExplosively(
>> FrameworkMethod.java:44)
>> org.junit.internal.runners.statements.RunBefores.
>> evaluate(RunBefores.java:24)
>> org.junit.internal.runners.statements.RunAfters.evaluate(
>> RunAfters.java:27)
>> org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>> org.junit.runners.Suite.runChild(Suite.java:127)
>> org.junit.runners.Suite.runChild(Suite.java:26)
>> org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>> org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>> org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>> [0m[ [0minfo [0m]  [0m [31mException encountered when attempting to run a
>> suite with class name: 
>> ittest.org.apache.geode.spark.connector.BasicIntegrationTest
>> *** ABORTED *** [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  org.apache.spark.SparkException: Only one
>> SparkContext may be running in this JVM (see SPARK-2243). To ignore this
>> error, set spark.driver.allowMultipleContexts = true. The currently
>> running SparkContext was created at: [0m [0m
>> [0m[ [0minfo [0m]  [0m 
>> [31morg.apache.spark.SparkContext.<init>(SparkContext.scala:80)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
>> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:50)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31morg.scalatest.BeforeAndAfterAll$class.
>> beforeAll(BeforeAndAfterAll.scala:187) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
>> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:30)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m 
>> [31morg.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mittest.org.apache.geode.spark.connector.
>> RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31morg.scalatest.tools.Framework.org
>> $scalatest$tools$Framework$$runSuite(Framework.scala:462) [0m [0m
>> [0m[ [0minfo [0m]  [0m 
>> [31morg.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:294)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:284)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
>> FutureTask.run(FutureTask.java:266) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
>> ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mjava.util.concurrent.
>> ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31mjava.lang.Thread.run(Thread.java:745) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
>> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1811)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
>> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1807)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at scala.Option.foreach(Option.scala:236)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
>> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$$anonfun$
>> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at scala.Option.foreach(Option.scala:236)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$.
>> assertNoOtherContextIsRunning(SparkContext.scala:1794) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at org.apache.spark.SparkContext$
>> .markPartiallyConstructed(SparkContext.scala:1833) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at 
>> org.apache.spark.SparkContext.<init>(SparkContext.scala:89)
>> [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  at ittest.org.apache.geode.spark.connector.
>> BasicIntegrationTest.beforeAll(BasicIntegrationTest.scala:58) [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m  ... [0m [0m
>> [0m[ [0minfo [0m]  [0mScalaTest [0m
>> [0m[ [0minfo [0m]  [0m [36mRun completed in 1 minute, 59 seconds. [0m [0m
>> [0m[ [0minfo [0m]  [0m [36mTotal number of tests run: 0 [0m [0m
>> [0m[ [0minfo [0m]  [0m [36mSuites: completed 1, aborted 3 [0m [0m
>> [0m[ [0minfo [0m]  [0m [36mTests: succeeded 0, failed 0, canceled 0,
>> ignored 0, pending 0 [0m [0m
>> [0m[ [0minfo [0m]  [0m [31m*** 3 SUITES ABORTED *** [0m [0m
>> [0m[ [31merror [0m]  [0mError: Total 3, Failed 0, Errors 3, Passed 0 [0m
>> [0m[ [31merror [0m]  [0mError during tests: [0m
>> [0m[ [31merror [0m]  [0m       
>> ittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest
>> [0m
>> [0m[ [31merror [0m]  [0m       
>> ittest.org.apache.geode.spark.connector.RetrieveRegionIntegrationTest
>> [0m
>> [0m[ [31merror [0m]  [0m       
>> ittest.org.apache.geode.spark.connector.BasicIntegrationTest
>> [0m
>> [0m[ [31merror [0m]  [0m(geode-spark-connector/it: [31mtest [0m)
>> sbt.TestsFailedException: Tests unsuccessful [0m
>> [0m[ [31merror [0m]  [0mTotal time: 128 s, completed Sep 23, 2016 4:12:09
>> PM [0m
>> Build step 'Execute shell' marked build as failure
>> Recording test results
>> Skipped archiving because build is not successful
>> 

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

Reply via email to