Yep, I’m seeing failures on any client app that doesn’t explicitly include spring as dependency.
Exception in thread "main" java.lang.NoClassDefFoundError: org/springframework/util/StringUtils at org.apache.geode.internal.net.SSLConfigurationFactory.configureSSLPropertiesFromSystemProperties(SSLConfigurationFactory.java:274) at org.apache.geode.internal.net.SSLConfigurationFactory.configureSSLPropertiesFromSystemProperties(SSLConfigurationFactory.java:270) at org.apache.geode.internal.net.SSLConfigurationFactory.createSSLConfigForComponent(SSLConfigurationFactory.java:138) at org.apache.geode.internal.net.SSLConfigurationFactory.getSSLConfigForComponent(SSLConfigurationFactory.java:67) at org.apache.geode.internal.net.SocketCreatorFactory.getSocketCreatorForComponent(SocketCreatorFactory.java:67) at org.apache.geode.distributed.internal.tcpserver.TcpClient.<init>(TcpClient.java:69) at org.apache.geode.cache.client.internal.AutoConnectionSourceImpl.<init>(AutoConnectionSourceImpl.java:114) at org.apache.geode.cache.client.internal.PoolImpl.getSourceImpl(PoolImpl.java:579) at org.apache.geode.cache.client.internal.PoolImpl.<init>(PoolImpl.java:219) at org.apache.geode.cache.client.internal.PoolImpl.create(PoolImpl.java:132) at org.apache.geode.internal.cache.PoolFactoryImpl.create(PoolFactoryImpl.java:319) at org.apache.geode.internal.cache.GemFireCacheImpl.determineDefaultPool(GemFireCacheImpl.java:2943) at org.apache.geode.internal.cache.GemFireCacheImpl.initializeDeclarativeCache(GemFireCacheImpl.java:1293) at org.apache.geode.internal.cache.GemFireCacheImpl.initialize(GemFireCacheImpl.java:1124) at org.apache.geode.internal.cache.GemFireCacheImpl.basicCreate(GemFireCacheImpl.java:765) at org.apache.geode.internal.cache.GemFireCacheImpl.createClient(GemFireCacheImpl.java:740) at org.apache.geode.cache.client.ClientCacheFactory.basicCreate(ClientCacheFactory.java:235) at org.apache.geode.cache.client.ClientCacheFactory.create(ClientCacheFactory.java:189) at HelloWorld.main(HelloWorld.java:25) Caused by: java.lang.ClassNotFoundException: org.springframework.util.StringUtils at java.net.URLClassLoader.findClass(URLClassLoader.java:381) at java.lang.ClassLoader.loadClass(ClassLoader.java:424) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331) at java.lang.ClassLoader.loadClass(ClassLoader.java:357) ... 19 more Anthony > On Sep 23, 2016, at 4:34 PM, Dan Smith <dsm...@pivotal.io> wrote: > > I created GEODE-1934 for this. It looks like the problem is actually that > our dependencies for geode-core are messed up. spring-core is marked > optional, but we're using it in critical places like this > SSLConfigurationFactory. > > In my opinion we shouldn't depend on spring-core at all unless we're > actually going to use it for things other than StringUtils. I think we've > accidentally introduced dependencies on it because the gfsh code in the > core is pulling in a bunch of spring libraries. > > -Dan > > > On Fri, Sep 23, 2016 at 9:12 AM, Apache Jenkins Server < > jenk...@builds.apache.org> wrote: > >> See <https://builds.apache.org/job/Geode-spark-connector/78/changes> >> >> Changes: >> >> [hkhamesra] GEODE-37 In spark connector we call TcpClient static method to >> get the >> >> [klund] GEODE-1906: fix misspelling of Successfully >> >> [upthewaterspout] GEODE-1915: Prevent deadlock registering instantiators >> with gateways >> >> ------------------------------------------ >> [...truncated 1883 lines...] >> 16/09/23 16:11:05 INFO HttpFileServer: HTTP File server directory is >> /tmp/spark-f13dac55-087f-4379-aeed-616fbdc7ffac/httpd- >> 02c1fab9-faa0-47f4-b0f3-fd44383eeeb3 >> 16/09/23 16:11:05 INFO HttpServer: Starting HTTP Server >> 16/09/23 16:11:05 INFO Utils: Successfully started service 'HTTP file >> server' on port 40135. >> 16/09/23 16:11:05 INFO SparkEnv: Registering OutputCommitCoordinator >> 16/09/23 16:11:10 WARN Utils: Service 'SparkUI' could not bind on port >> 4040. Attempting port 4041. >> 16/09/23 16:11:15 INFO Utils: Successfully started service 'SparkUI' on >> port 4041. >> 16/09/23 16:11:15 INFO SparkUI: Started SparkUI at http://localhost:4041 >> 16/09/23 16:11:15 INFO Executor: Starting executor ID <driver> on host >> localhost >> 16/09/23 16:11:15 INFO AkkaUtils: Connecting to HeartbeatReceiver: >> akka.tcp://sparkDriver@localhost:54872/user/HeartbeatReceiver >> 16/09/23 16:11:16 INFO NettyBlockTransferService: Server created on 41182 >> 16/09/23 16:11:16 INFO BlockManagerMaster: Trying to register BlockManager >> 16/09/23 16:11:16 INFO BlockManagerMasterActor: Registering block manager >> localhost:41182 with 2.8 GB RAM, BlockManagerId(<driver>, localhost, 41182) >> 16/09/23 16:11:16 INFO BlockManagerMaster: Registered BlockManager >> === GeodeRunner: stop server 1. >> === GeodeRunner: stop server 2. >> [0m[ [0minfo [0m] [0m [32mRetrieveRegionIntegrationTest: [0m [0m >> ...... >> >> === GeodeRunner: stop locator >> ... >> Successfully stop Geode locator at port 27662. >> === GeodeRunner: starting locator on port 23825 >> === GeodeRunner: waiting for locator on port 23825 >> ....=== GeodeRunner: done waiting for locator on port 23825 >> === GeodeRunner: starting server1 with clientPort 28993 >> === GeodeRunner: starting server2 with clientPort 26318 >> === GeodeRunner: starting server3 with clientPort 29777 >> === GeodeRunner: starting server4 with clientPort 22946 >> .... >> ............................................Locator in >> /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >> connector/geode-spark-connector/target/testgeode/locator on >> hemera.apache.org[23825] as locator is currently online. >> Process ID: 1860 >> Uptime: 4 seconds >> GemFire Version: 1.0.0-incubating-SNAPSHOT >> Java Version: 1.8.0_66 >> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >> locator/locator.log >> JVM Arguments: -Dgemfire.enable-cluster-configuration=true >> -Dgemfire.load-cluster-configuration-from-dir=false >> -Dgemfire.jmx-manager-http-port=29684 >> -Dgemfire.launcher.registerSignalHandlers=true >> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >> gcInterval=9223372036854775806 >> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0- >> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/ >> lib/geode-dependencies.jar >> >> Successfully connected to: JMX Manager [host=hemera.apache.org, port=1099] >> >> Cluster configuration service is up and running. >> >> ................ >> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server4 >> on hemera.apache.org[22946] as server4 is currently online. >> Process ID: 2204 >> Uptime: 8 seconds >> GemFire Version: 1.0.0-incubating-SNAPSHOT >> Java Version: 1.8.0_66 >> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >> server4/server4.log >> JVM Arguments: -Dgemfire.locators=localhost[23825] >> -Dgemfire.use-cluster-configuration=true >> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml >> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >> -XX:OnOutOfMemoryError=kill -KILL %p >> -Dgemfire.launcher.registerSignalHandlers=true >> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >> gcInterval=9223372036854775806 >> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0- >> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/ >> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >> spark-connector/geode-assembly/build/install/apache- >> geode/lib/geode-dependencies.jar >> >> .. >> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1 >> on hemera.apache.org[28993] as server1 is currently online. >> Process ID: 2199 >> Uptime: 8 seconds >> GemFire Version: 1.0.0-incubating-SNAPSHOT >> Java Version: 1.8.0_66 >> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >> server1/server1.log >> JVM Arguments: -Dgemfire.locators=localhost[23825] >> -Dgemfire.use-cluster-configuration=true >> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml >> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >> -XX:OnOutOfMemoryError=kill -KILL %p >> -Dgemfire.launcher.registerSignalHandlers=true >> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >> gcInterval=9223372036854775806 >> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0- >> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/ >> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >> spark-connector/geode-assembly/build/install/apache- >> geode/lib/geode-dependencies.jar >> >> >> >> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2 >> on hemera.apache.org[26318] as server2 is currently online. >> Process ID: 2153 >> Uptime: 9 seconds >> GemFire Version: 1.0.0-incubating-SNAPSHOT >> Java Version: 1.8.0_66 >> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >> server2/server2.log >> JVM Arguments: -Dgemfire.locators=localhost[23825] >> -Dgemfire.use-cluster-configuration=true >> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml >> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >> -XX:OnOutOfMemoryError=kill -KILL %p >> -Dgemfire.launcher.registerSignalHandlers=true >> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >> gcInterval=9223372036854775806 >> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0- >> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/ >> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >> spark-connector/geode-assembly/build/install/apache- >> geode/lib/geode-dependencies.jar >> >> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server3 >> on hemera.apache.org[29777] as server3 is currently online. >> Process ID: 2175 >> Uptime: 9 seconds >> GemFire Version: 1.0.0-incubating-SNAPSHOT >> Java Version: 1.8.0_66 >> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >> server3/server3.log >> JVM Arguments: -Dgemfire.locators=localhost[23825] >> -Dgemfire.use-cluster-configuration=true >> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >> connector/geode-spark-connector/src/it/resources/test-retrieve-regions.xml >> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >> -XX:OnOutOfMemoryError=kill -KILL %p >> -Dgemfire.launcher.registerSignalHandlers=true >> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >> gcInterval=9223372036854775806 >> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0- >> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/ >> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >> spark-connector/geode-assembly/build/install/apache- >> geode/lib/geode-dependencies.jar >> >> All WrappedArray(28993, 26318, 29777, 22946).length servers have been >> started >> Deploying:geode-functions_2.10-0.5.0.jar >> 16/09/23 16:11:43 WARN SparkContext: Another SparkContext is being >> constructed (or threw an exception in its constructor). This may indicate >> an error, since only one SparkContext may be running in this JVM (see >> SPARK-2243). The other SparkContext was created at: >> org.apache.spark.api.java.JavaSparkContext.<init>( >> JavaSparkContext.scala:61) >> ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest. >> setUpBeforeClass(JavaApiIntegrationTest.java:75) >> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java: >> 62) >> sun.reflect.DelegatingMethodAccessorImpl.invoke( >> DelegatingMethodAccessorImpl.java:43) >> java.lang.reflect.Method.invoke(Method.java:497) >> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall( >> FrameworkMethod.java:47) >> org.junit.internal.runners.model.ReflectiveCallable.run( >> ReflectiveCallable.java:12) >> org.junit.runners.model.FrameworkMethod.invokeExplosively( >> FrameworkMethod.java:44) >> org.junit.internal.runners.statements.RunBefores. >> evaluate(RunBefores.java:24) >> org.junit.internal.runners.statements.RunAfters.evaluate( >> RunAfters.java:27) >> org.junit.runners.ParentRunner.run(ParentRunner.java:309) >> org.junit.runners.Suite.runChild(Suite.java:127) >> org.junit.runners.Suite.runChild(Suite.java:26) >> org.junit.runners.ParentRunner$3.run(ParentRunner.java:238) >> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63) >> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236) >> org.junit.runners.ParentRunner.access$000(ParentRunner.java:53) >> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229) >> org.junit.runners.ParentRunner.run(ParentRunner.java:309) >> [0m[ [0minfo [0m] [0m [31mException encountered when attempting to run a >> suite with class name: >> ittest.org.apache.geode.spark.connector.RetrieveRegionIntegrationTest >> *** ABORTED *** [0m [0m >> [0m[ [0minfo [0m] [0m [31m org.apache.spark.SparkException: Only one >> SparkContext may be running in this JVM (see SPARK-2243). To ignore this >> error, set spark.driver.allowMultipleContexts = true. The currently >> running SparkContext was created at: [0m [0m >> [0m[ [0minfo [0m] [0m >> [31morg.apache.spark.SparkContext.<init>(SparkContext.scala:80) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:50) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31morg.scalatest.BeforeAndAfterAll$class. >> beforeAll(BeforeAndAfterAll.scala:187) [0m [0m >> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:30) >> [0m [0m >> [0m[ [0minfo [0m] [0m >> [31morg.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >> RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31morg.scalatest.tools.Framework.org >> $scalatest$tools$Framework$$runSuite(Framework.scala:462) [0m [0m >> [0m[ [0minfo [0m] [0m >> [31morg.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:294) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:284) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >> FutureTask.run(FutureTask.java:266) [0m [0m >> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >> ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [0m [0m >> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >> ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [0m [0m >> [0m[ [0minfo [0m] [0m [31mjava.lang.Thread.run(Thread.java:745) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1811) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1807) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at scala.Option.foreach(Option.scala:236) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at scala.Option.foreach(Option.scala:236) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$. >> assertNoOtherContextIsRunning(SparkContext.scala:1794) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$ >> .markPartiallyConstructed(SparkContext.scala:1833) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at >> org.apache.spark.SparkContext.<init>(SparkContext.scala:89) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at ittest.org.apache.geode.spark.connector. >> RetrieveRegionIntegrationTest.beforeAll(RetrieveRegionIntegrationTest.scala:51) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m ... [0m [0m >> [0m[ [0minfo [0m] [0m [32mBasicIntegrationTest: [0m [0m >> === GeodeRunner: stop server 1. >> === GeodeRunner: stop server 2. >> === GeodeRunner: stop server 3. >> === GeodeRunner: stop server 4. >> ............ >> >> >> >> === GeodeRunner: stop locator >> ... >> Successfully stop Geode locator at port 23825. >> === GeodeRunner: starting locator on port 23573 >> === GeodeRunner: waiting for locator on port 23573 >> ....=== GeodeRunner: done waiting for locator on port 23573 >> === GeodeRunner: starting server1 with clientPort 27897 >> === GeodeRunner: starting server2 with clientPort 20289 >> .... >> ....................Locator in /x1/jenkins/jenkins-slave/ >> workspace/Geode-spark-connector/geode-spark-connector/geode-spark- >> connector/target/testgeode/locator on hemera.apache.org[23573] as locator >> is currently online. >> Process ID: 3273 >> Uptime: 4 seconds >> GemFire Version: 1.0.0-incubating-SNAPSHOT >> Java Version: 1.8.0_66 >> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >> locator/locator.log >> JVM Arguments: -Dgemfire.enable-cluster-configuration=true >> -Dgemfire.load-cluster-configuration-from-dir=false >> -Dgemfire.jmx-manager-http-port=23053 >> -Dgemfire.launcher.registerSignalHandlers=true >> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >> gcInterval=9223372036854775806 >> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0- >> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/ >> lib/geode-dependencies.jar >> >> Successfully connected to: JMX Manager [host=hemera.apache.org, port=1099] >> >> Cluster configuration service is up and running. >> >> ........ >> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server2 >> on hemera.apache.org[20289] as server2 is currently online. >> Process ID: 3465 >> Uptime: 7 seconds >> GemFire Version: 1.0.0-incubating-SNAPSHOT >> Java Version: 1.8.0_66 >> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >> server2/server2.log >> JVM Arguments: -Dgemfire.locators=localhost[23573] >> -Dgemfire.use-cluster-configuration=true >> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >> connector/geode-spark-connector/src/it/resources/test-regions.xml >> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >> -XX:OnOutOfMemoryError=kill -KILL %p >> -Dgemfire.launcher.registerSignalHandlers=true >> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >> gcInterval=9223372036854775806 >> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0- >> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/ >> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >> spark-connector/geode-assembly/build/install/apache- >> geode/lib/geode-dependencies.jar >> >> >> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/server1 >> on hemera.apache.org[27897] as server1 is currently online. >> Process ID: 3505 >> Uptime: 7 seconds >> GemFire Version: 1.0.0-incubating-SNAPSHOT >> Java Version: 1.8.0_66 >> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >> server1/server1.log >> JVM Arguments: -Dgemfire.locators=localhost[23573] >> -Dgemfire.use-cluster-configuration=true >> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >> connector/geode-spark-connector/src/it/resources/test-regions.xml >> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >> -XX:OnOutOfMemoryError=kill -KILL %p >> -Dgemfire.launcher.registerSignalHandlers=true >> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >> gcInterval=9223372036854775806 >> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-assembly/build/install/apache-geode/lib/geode-core-1.0.0- >> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Geode-spark- >> connector/geode-spark-connector/geode-spark-connector/./target/scala-2.10/ >> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >> spark-connector/geode-assembly/build/install/apache- >> geode/lib/geode-dependencies.jar >> >> All WrappedArray(27897, 20289).length servers have been started >> Deploying:geode-functions_2.10-0.5.0.jar >> 16/09/23 16:12:09 WARN SparkContext: Another SparkContext is being >> constructed (or threw an exception in its constructor). This may indicate >> an error, since only one SparkContext may be running in this JVM (see >> SPARK-2243). The other SparkContext was created at: >> org.apache.spark.api.java.JavaSparkContext.<init>( >> JavaSparkContext.scala:61) >> ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest. >> setUpBeforeClass(JavaApiIntegrationTest.java:75) >> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java: >> 62) >> sun.reflect.DelegatingMethodAccessorImpl.invoke( >> DelegatingMethodAccessorImpl.java:43) >> java.lang.reflect.Method.invoke(Method.java:497) >> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall( >> FrameworkMethod.java:47) >> org.junit.internal.runners.model.ReflectiveCallable.run( >> ReflectiveCallable.java:12) >> org.junit.runners.model.FrameworkMethod.invokeExplosively( >> FrameworkMethod.java:44) >> org.junit.internal.runners.statements.RunBefores. >> evaluate(RunBefores.java:24) >> org.junit.internal.runners.statements.RunAfters.evaluate( >> RunAfters.java:27) >> org.junit.runners.ParentRunner.run(ParentRunner.java:309) >> org.junit.runners.Suite.runChild(Suite.java:127) >> org.junit.runners.Suite.runChild(Suite.java:26) >> org.junit.runners.ParentRunner$3.run(ParentRunner.java:238) >> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63) >> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236) >> org.junit.runners.ParentRunner.access$000(ParentRunner.java:53) >> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229) >> org.junit.runners.ParentRunner.run(ParentRunner.java:309) >> [0m[ [0minfo [0m] [0m [31mException encountered when attempting to run a >> suite with class name: >> ittest.org.apache.geode.spark.connector.BasicIntegrationTest >> *** ABORTED *** [0m [0m >> [0m[ [0minfo [0m] [0m [31m org.apache.spark.SparkException: Only one >> SparkContext may be running in this JVM (see SPARK-2243). To ignore this >> error, set spark.driver.allowMultipleContexts = true. The currently >> running SparkContext was created at: [0m [0m >> [0m[ [0minfo [0m] [0m >> [31morg.apache.spark.SparkContext.<init>(SparkContext.scala:80) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:50) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31morg.scalatest.BeforeAndAfterAll$class. >> beforeAll(BeforeAndAfterAll.scala:187) [0m [0m >> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrationTest.scala:30) >> [0m [0m >> [0m[ [0minfo [0m] [0m >> [31morg.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >> RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31morg.scalatest.tools.Framework.org >> $scalatest$tools$Framework$$runSuite(Framework.scala:462) [0m [0m >> [0m[ [0minfo [0m] [0m >> [31morg.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:294) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:284) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >> FutureTask.run(FutureTask.java:266) [0m [0m >> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >> ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [0m [0m >> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >> ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [0m [0m >> [0m[ [0minfo [0m] [0m [31mjava.lang.Thread.run(Thread.java:745) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1811) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(SparkContext.scala:1807) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at scala.Option.foreach(Option.scala:236) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at scala.Option.foreach(Option.scala:236) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$. >> assertNoOtherContextIsRunning(SparkContext.scala:1794) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$ >> .markPartiallyConstructed(SparkContext.scala:1833) [0m [0m >> [0m[ [0minfo [0m] [0m [31m at >> org.apache.spark.SparkContext.<init>(SparkContext.scala:89) >> [0m [0m >> [0m[ [0minfo [0m] [0m [31m at ittest.org.apache.geode.spark.connector. >> BasicIntegrationTest.beforeAll(BasicIntegrationTest.scala:58) [0m [0m >> [0m[ [0minfo [0m] [0m [31m ... [0m [0m >> [0m[ [0minfo [0m] [0mScalaTest [0m >> [0m[ [0minfo [0m] [0m [36mRun completed in 1 minute, 59 seconds. [0m [0m >> [0m[ [0minfo [0m] [0m [36mTotal number of tests run: 0 [0m [0m >> [0m[ [0minfo [0m] [0m [36mSuites: completed 1, aborted 3 [0m [0m >> [0m[ [0minfo [0m] [0m [36mTests: succeeded 0, failed 0, canceled 0, >> ignored 0, pending 0 [0m [0m >> [0m[ [0minfo [0m] [0m [31m*** 3 SUITES ABORTED *** [0m [0m >> [0m[ [31merror [0m] [0mError: Total 3, Failed 0, Errors 3, Passed 0 [0m >> [0m[ [31merror [0m] [0mError during tests: [0m >> [0m[ [31merror [0m] [0m >> ittest.org.apache.geode.spark.connector.RDDJoinRegionIntegrationTest >> [0m >> [0m[ [31merror [0m] [0m >> ittest.org.apache.geode.spark.connector.RetrieveRegionIntegrationTest >> [0m >> [0m[ [31merror [0m] [0m >> ittest.org.apache.geode.spark.connector.BasicIntegrationTest >> [0m >> [0m[ [31merror [0m] [0m(geode-spark-connector/it: [31mtest [0m) >> sbt.TestsFailedException: Tests unsuccessful [0m >> [0m[ [31merror [0m] [0mTotal time: 128 s, completed Sep 23, 2016 4:12:09 >> PM [0m >> Build step 'Execute shell' marked build as failure >> Recording test results >> Skipped archiving because build is not successful >>
signature.asc
Description: Message signed with OpenPGP using GPGMail