org.apache.geode.internal.lang.StringUtils includes isEmpty(String) -Kirk
On Friday, September 23, 2016, Udo Kohlmeyer <ukohlme...@pivotal.io> wrote: > I can easily fix this. > > Sure we have a utility lying around in the core that can handle > "String.isEmpty" > > --Udo > > > On 24/09/2016 9:56 AM, Anthony Baker wrote: > >> Yep, I’m seeing failures on any client app that doesn’t explicitly >> include spring as dependency. >> >> Exception in thread "main" java.lang.NoClassDefFoundError: >> org/springframework/util/StringUtils >> at org.apache.geode.internal.net.SSLConfigurationFactory.config >> ureSSLPropertiesFromSystemProperties(SSLConfigurationFactory.java:274) >> at org.apache.geode.internal.net.SSLConfigurationFactory.config >> ureSSLPropertiesFromSystemProperties(SSLConfigurationFactory.java:270) >> at org.apache.geode.internal.net.SSLConfigurationFactory.create >> SSLConfigForComponent(SSLConfigurationFactory.java:138) >> at org.apache.geode.internal.net.SSLConfigurationFactory.getSSL >> ConfigForComponent(SSLConfigurationFactory.java:67) >> at org.apache.geode.internal.net.SocketCreatorFactory.getSocket >> CreatorForComponent(SocketCreatorFactory.java:67) >> at org.apache.geode.distributed.internal.tcpserver.TcpClient.<i >> nit>(TcpClient.java:69) >> at org.apache.geode.cache.client.internal.AutoConnectionSourceI >> mpl.<init>(AutoConnectionSourceImpl.java:114) >> at org.apache.geode.cache.client.internal.PoolImpl.getSourceImp >> l(PoolImpl.java:579) >> at org.apache.geode.cache.client.internal.PoolImpl.<init>(PoolI >> mpl.java:219) >> at org.apache.geode.cache.client.internal.PoolImpl.create(PoolI >> mpl.java:132) >> at org.apache.geode.internal.cache.PoolFactoryImpl.create(PoolF >> actoryImpl.java:319) >> at org.apache.geode.internal.cache.GemFireCacheImpl.determineDe >> faultPool(GemFireCacheImpl.java:2943) >> at org.apache.geode.internal.cache.GemFireCacheImpl.initializeD >> eclarativeCache(GemFireCacheImpl.java:1293) >> at org.apache.geode.internal.cache.GemFireCacheImpl.initialize( >> GemFireCacheImpl.java:1124) >> at org.apache.geode.internal.cache.GemFireCacheImpl.basicCreate >> (GemFireCacheImpl.java:765) >> at org.apache.geode.internal.cache.GemFireCacheImpl.createClien >> t(GemFireCacheImpl.java:740) >> at org.apache.geode.cache.client.ClientCacheFactory.basicCreate >> (ClientCacheFactory.java:235) >> at org.apache.geode.cache.client.ClientCacheFactory.create(Clie >> ntCacheFactory.java:189) >> at HelloWorld.main(HelloWorld.java:25) >> Caused by: java.lang.ClassNotFoundException: >> org.springframework.util.StringUtils >> at java.net.URLClassLoader.findClass(URLClassLoader.java:381) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >> ... 19 more >> >> Anthony >> >> >> On Sep 23, 2016, at 4:34 PM, Dan Smith <dsm...@pivotal.io> wrote: >>> >>> I created GEODE-1934 for this. It looks like the problem is actually that >>> our dependencies for geode-core are messed up. spring-core is marked >>> optional, but we're using it in critical places like this >>> SSLConfigurationFactory. >>> >>> In my opinion we shouldn't depend on spring-core at all unless we're >>> actually going to use it for things other than StringUtils. I think we've >>> accidentally introduced dependencies on it because the gfsh code in the >>> core is pulling in a bunch of spring libraries. >>> >>> -Dan >>> >>> >>> On Fri, Sep 23, 2016 at 9:12 AM, Apache Jenkins Server < >>> jenk...@builds.apache.org> wrote: >>> >>> See <https://builds.apache.org/job/Geode-spark-connector/78/changes> >>>> >>>> Changes: >>>> >>>> [hkhamesra] GEODE-37 In spark connector we call TcpClient static method >>>> to >>>> get the >>>> >>>> [klund] GEODE-1906: fix misspelling of Successfully >>>> >>>> [upthewaterspout] GEODE-1915: Prevent deadlock registering instantiators >>>> with gateways >>>> >>>> ------------------------------------------ >>>> [...truncated 1883 lines...] >>>> 16/09/23 16:11:05 INFO HttpFileServer: HTTP File server directory is >>>> /tmp/spark-f13dac55-087f-4379-aeed-616fbdc7ffac/httpd- >>>> 02c1fab9-faa0-47f4-b0f3-fd44383eeeb3 >>>> 16/09/23 16:11:05 INFO HttpServer: Starting HTTP Server >>>> 16/09/23 16:11:05 INFO Utils: Successfully started service 'HTTP file >>>> server' on port 40135. >>>> 16/09/23 16:11:05 INFO SparkEnv: Registering OutputCommitCoordinator >>>> 16/09/23 16:11:10 WARN Utils: Service 'SparkUI' could not bind on port >>>> 4040. Attempting port 4041. >>>> 16/09/23 16:11:15 INFO Utils: Successfully started service 'SparkUI' on >>>> port 4041. >>>> 16/09/23 16:11:15 INFO SparkUI: Started SparkUI at >>>> http://localhost:4041 >>>> 16/09/23 16:11:15 INFO Executor: Starting executor ID <driver> on host >>>> localhost >>>> 16/09/23 16:11:15 INFO AkkaUtils: Connecting to HeartbeatReceiver: >>>> akka.tcp://sparkDriver@localhost:54872/user/HeartbeatReceiver >>>> 16/09/23 16:11:16 INFO NettyBlockTransferService: Server created on >>>> 41182 >>>> 16/09/23 16:11:16 INFO BlockManagerMaster: Trying to register >>>> BlockManager >>>> 16/09/23 16:11:16 INFO BlockManagerMasterActor: Registering block >>>> manager >>>> localhost:41182 with 2.8 GB RAM, BlockManagerId(<driver>, localhost, >>>> 41182) >>>> 16/09/23 16:11:16 INFO BlockManagerMaster: Registered BlockManager >>>> === GeodeRunner: stop server 1. >>>> === GeodeRunner: stop server 2. >>>> [0m[ [0minfo [0m] [0m [32mRetrieveRegionIntegrationTest: [0m [0m >>>> ...... >>>> >>>> === GeodeRunner: stop locator >>>> ... >>>> Successfully stop Geode locator at port 27662. >>>> === GeodeRunner: starting locator on port 23825 >>>> === GeodeRunner: waiting for locator on port 23825 >>>> ....=== GeodeRunner: done waiting for locator on port 23825 >>>> === GeodeRunner: starting server1 with clientPort 28993 >>>> === GeodeRunner: starting server2 with clientPort 26318 >>>> === GeodeRunner: starting server3 with clientPort 29777 >>>> === GeodeRunner: starting server4 with clientPort 22946 >>>> .... >>>> ............................................Locator in >>>> /x1/jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >>>> connector/geode-spark-connector/target/testgeode/locator on >>>> hemera.apache.org[23825] as locator is currently online. >>>> Process ID: 1860 >>>> Uptime: 4 seconds >>>> GemFire Version: 1.0.0-incubating-SNAPSHOT >>>> Java Version: 1.8.0_66 >>>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >>>> locator/locator.log >>>> JVM Arguments: -Dgemfire.enable-cluster-configuration=true >>>> -Dgemfire.load-cluster-configuration-from-dir=false >>>> -Dgemfire.jmx-manager-http-port=29684 -Dgemfire.launcher.registerSig >>>> nalHandlers=true >>>> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >>>> gcInterval=9223372036854775806 >>>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/lib/ >>>> geode-core-1.0.0- >>>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/ >>>> Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/ >>>> lib/geode-dependencies.jar >>>> >>>> Successfully connected to: JMX Manager [host=hemera.apache.org, >>>> port=1099] >>>> >>>> Cluster configuration service is up and running. >>>> >>>> ................ >>>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/ >>>> target/testgeode/server4 >>>> on hemera.apache.org[22946] as server4 is currently online. >>>> Process ID: 2204 >>>> Uptime: 8 seconds >>>> GemFire Version: 1.0.0-incubating-SNAPSHOT >>>> Java Version: 1.8.0_66 >>>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >>>> server4/server4.log >>>> JVM Arguments: -Dgemfire.locators=localhost[23825] >>>> -Dgemfire.use-cluster-configuration=true >>>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >>>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >>>> connector/geode-spark-connector/src/it/resources/test- >>>> retrieve-regions.xml >>>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >>>> -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSig >>>> nalHandlers=true >>>> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >>>> gcInterval=9223372036854775806 >>>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/lib/ >>>> geode-core-1.0.0- >>>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/ >>>> Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/./ >>>> target/scala-2.10/ >>>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >>>> spark-connector/geode-assembly/build/install/apache- >>>> geode/lib/geode-dependencies.jar >>>> >>>> .. >>>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/ >>>> target/testgeode/server1 >>>> on hemera.apache.org[28993] as server1 is currently online. >>>> Process ID: 2199 >>>> Uptime: 8 seconds >>>> GemFire Version: 1.0.0-incubating-SNAPSHOT >>>> Java Version: 1.8.0_66 >>>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >>>> server1/server1.log >>>> JVM Arguments: -Dgemfire.locators=localhost[23825] >>>> -Dgemfire.use-cluster-configuration=true >>>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >>>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >>>> connector/geode-spark-connector/src/it/resources/test- >>>> retrieve-regions.xml >>>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >>>> -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSig >>>> nalHandlers=true >>>> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >>>> gcInterval=9223372036854775806 >>>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/lib/ >>>> geode-core-1.0.0- >>>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/ >>>> Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/./ >>>> target/scala-2.10/ >>>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >>>> spark-connector/geode-assembly/build/install/apache- >>>> geode/lib/geode-dependencies.jar >>>> >>>> >>>> >>>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/ >>>> target/testgeode/server2 >>>> on hemera.apache.org[26318] as server2 is currently online. >>>> Process ID: 2153 >>>> Uptime: 9 seconds >>>> GemFire Version: 1.0.0-incubating-SNAPSHOT >>>> Java Version: 1.8.0_66 >>>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >>>> server2/server2.log >>>> JVM Arguments: -Dgemfire.locators=localhost[23825] >>>> -Dgemfire.use-cluster-configuration=true >>>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >>>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >>>> connector/geode-spark-connector/src/it/resources/test- >>>> retrieve-regions.xml >>>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >>>> -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSig >>>> nalHandlers=true >>>> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >>>> gcInterval=9223372036854775806 >>>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/lib/ >>>> geode-core-1.0.0- >>>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/ >>>> Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/./ >>>> target/scala-2.10/ >>>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >>>> spark-connector/geode-assembly/build/install/apache- >>>> geode/lib/geode-dependencies.jar >>>> >>>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/ >>>> target/testgeode/server3 >>>> on hemera.apache.org[29777] as server3 is currently online. >>>> Process ID: 2175 >>>> Uptime: 9 seconds >>>> GemFire Version: 1.0.0-incubating-SNAPSHOT >>>> Java Version: 1.8.0_66 >>>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >>>> server3/server3.log >>>> JVM Arguments: -Dgemfire.locators=localhost[23825] >>>> -Dgemfire.use-cluster-configuration=true >>>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >>>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >>>> connector/geode-spark-connector/src/it/resources/test- >>>> retrieve-regions.xml >>>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >>>> -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSig >>>> nalHandlers=true >>>> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >>>> gcInterval=9223372036854775806 >>>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/lib/ >>>> geode-core-1.0.0- >>>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/ >>>> Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/./ >>>> target/scala-2.10/ >>>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >>>> spark-connector/geode-assembly/build/install/apache- >>>> geode/lib/geode-dependencies.jar >>>> >>>> All WrappedArray(28993, 26318, 29777, 22946).length servers have been >>>> started >>>> Deploying:geode-functions_2.10-0.5.0.jar >>>> 16/09/23 16:11:43 WARN SparkContext: Another SparkContext is being >>>> constructed (or threw an exception in its constructor). This may >>>> indicate >>>> an error, since only one SparkContext may be running in this JVM (see >>>> SPARK-2243). The other SparkContext was created at: >>>> org.apache.spark.api.java.JavaSparkContext.<init>( >>>> JavaSparkContext.scala:61) >>>> ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest. >>>> setUpBeforeClass(JavaApiIntegrationTest.java:75) >>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >>>> ssorImpl.java: >>>> 62) >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke( >>>> DelegatingMethodAccessorImpl.java:43) >>>> java.lang.reflect.Method.invoke(Method.java:497) >>>> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall( >>>> FrameworkMethod.java:47) >>>> org.junit.internal.runners.model.ReflectiveCallable.run( >>>> ReflectiveCallable.java:12) >>>> org.junit.runners.model.FrameworkMethod.invokeExplosively( >>>> FrameworkMethod.java:44) >>>> org.junit.internal.runners.statements.RunBefores. >>>> evaluate(RunBefores.java:24) >>>> org.junit.internal.runners.statements.RunAfters.evaluate( >>>> RunAfters.java:27) >>>> org.junit.runners.ParentRunner.run(ParentRunner.java:309) >>>> org.junit.runners.Suite.runChild(Suite.java:127) >>>> org.junit.runners.Suite.runChild(Suite.java:26) >>>> org.junit.runners.ParentRunner$3.run(ParentRunner.java:238) >>>> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63) >>>> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236) >>>> org.junit.runners.ParentRunner.access$000(ParentRunner.java:53) >>>> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229) >>>> org.junit.runners.ParentRunner.run(ParentRunner.java:309) >>>> [0m[ [0minfo [0m] [0m [31mException encountered when attempting to run >>>> a >>>> suite with class name: ittest.org.apache.geode.spark. >>>> connector.RetrieveRegionIntegrationTest >>>> *** ABORTED *** [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m org.apache.spark.SparkException: Only one >>>> SparkContext may be running in this JVM (see SPARK-2243). To ignore this >>>> error, set spark.driver.allowMultipleContexts = true. The currently >>>> running SparkContext was created at: [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.apache.spark.SparkCont >>>> ext.<init>(SparkContext.scala:80) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >>>> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrat >>>> ionTest.scala:50) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.scalatest.BeforeAndAfterAll$class. >>>> beforeAll(BeforeAndAfterAll.scala:187) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >>>> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrat >>>> ionTest.scala:30) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.scalatest.BeforeAndAft >>>> erAll$class.run(BeforeAndAfterAll.scala:253) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >>>> RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.scalatest.tools.Framework.org >>>> $scalatest$tools$Framework$$runSuite(Framework.scala:462) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.scalatest.tools.Framew >>>> ork$ScalaTestTask.execute(Framework.scala:671) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:294) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:284) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >>>> FutureTask.run(FutureTask.java:266) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >>>> ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >>>> ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mjava.lang.Thread.run(Thread.java:745) [0m >>>> [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >>>> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(Spar >>>> kContext.scala:1811) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >>>> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(Spar >>>> kContext.scala:1807) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at scala.Option.foreach(Option.scala:236) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >>>> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >>>> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at scala.Option.foreach(Option.scala:236) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$. >>>> assertNoOtherContextIsRunning(SparkContext.scala:1794) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$ >>>> .markPartiallyConstructed(SparkContext.scala:1833) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext. >>>> <init>(SparkContext.scala:89) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at ittest.org.apache.geode.spark. >>>> connector. >>>> RetrieveRegionIntegrationTest.beforeAll(RetrieveRegionIntegr >>>> ationTest.scala:51) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m ... [0m [0m >>>> [0m[ [0minfo [0m] [0m [32mBasicIntegrationTest: [0m [0m >>>> === GeodeRunner: stop server 1. >>>> === GeodeRunner: stop server 2. >>>> === GeodeRunner: stop server 3. >>>> === GeodeRunner: stop server 4. >>>> ............ >>>> >>>> >>>> >>>> === GeodeRunner: stop locator >>>> ... >>>> Successfully stop Geode locator at port 23825. >>>> === GeodeRunner: starting locator on port 23573 >>>> === GeodeRunner: waiting for locator on port 23573 >>>> ....=== GeodeRunner: done waiting for locator on port 23573 >>>> === GeodeRunner: starting server1 with clientPort 27897 >>>> === GeodeRunner: starting server2 with clientPort 20289 >>>> .... >>>> ....................Locator in /x1/jenkins/jenkins-slave/ >>>> workspace/Geode-spark-connector/geode-spark-connector/geode-spark- >>>> connector/target/testgeode/locator on hemera.apache.org[23573] as >>>> locator >>>> is currently online. >>>> Process ID: 3273 >>>> Uptime: 4 seconds >>>> GemFire Version: 1.0.0-incubating-SNAPSHOT >>>> Java Version: 1.8.0_66 >>>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >>>> locator/locator.log >>>> JVM Arguments: -Dgemfire.enable-cluster-configuration=true >>>> -Dgemfire.load-cluster-configuration-from-dir=false >>>> -Dgemfire.jmx-manager-http-port=23053 -Dgemfire.launcher.registerSig >>>> nalHandlers=true >>>> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >>>> gcInterval=9223372036854775806 >>>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/lib/ >>>> geode-core-1.0.0- >>>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/ >>>> Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/ >>>> lib/geode-dependencies.jar >>>> >>>> Successfully connected to: JMX Manager [host=hemera.apache.org, >>>> port=1099] >>>> >>>> Cluster configuration service is up and running. >>>> >>>> ........ >>>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/ >>>> target/testgeode/server2 >>>> on hemera.apache.org[20289] as server2 is currently online. >>>> Process ID: 3465 >>>> Uptime: 7 seconds >>>> GemFire Version: 1.0.0-incubating-SNAPSHOT >>>> Java Version: 1.8.0_66 >>>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >>>> server2/server2.log >>>> JVM Arguments: -Dgemfire.locators=localhost[23573] >>>> -Dgemfire.use-cluster-configuration=true >>>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >>>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >>>> connector/geode-spark-connector/src/it/resources/test-regions.xml >>>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >>>> -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSig >>>> nalHandlers=true >>>> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >>>> gcInterval=9223372036854775806 >>>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/lib/ >>>> geode-core-1.0.0- >>>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/ >>>> Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/./ >>>> target/scala-2.10/ >>>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >>>> spark-connector/geode-assembly/build/install/apache- >>>> geode/lib/geode-dependencies.jar >>>> >>>> >>>> Server in /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/ >>>> target/testgeode/server1 >>>> on hemera.apache.org[27897] as server1 is currently online. >>>> Process ID: 3505 >>>> Uptime: 7 seconds >>>> GemFire Version: 1.0.0-incubating-SNAPSHOT >>>> Java Version: 1.8.0_66 >>>> Log File: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/target/testgeode/ >>>> server1/server1.log >>>> JVM Arguments: -Dgemfire.locators=localhost[23573] >>>> -Dgemfire.use-cluster-configuration=true >>>> -Dgemfire.bind-address=localhost -Dgemfire.cache-xml-file=/x1/ >>>> jenkins/jenkins-slave/workspace/Geode-spark-connector/geode-spark- >>>> connector/geode-spark-connector/src/it/resources/test-regions.xml >>>> -Dgemfire.http-service-port=8080 -Dgemfire.start-dev-rest-api=false >>>> -XX:OnOutOfMemoryError=kill -KILL %p -Dgemfire.launcher.registerSig >>>> nalHandlers=true >>>> -Djava.awt.headless=true -Dsun.rmi.dgc.server. >>>> gcInterval=9223372036854775806 >>>> Class-Path: /x1/jenkins/jenkins-slave/workspace/Geode-spark- >>>> connector/geode-assembly/build/install/apache-geode/lib/ >>>> geode-core-1.0.0- >>>> incubating-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/ >>>> Geode-spark- >>>> connector/geode-spark-connector/geode-spark-connector/./ >>>> target/scala-2.10/ >>>> it-classes:/x1/jenkins/jenkins-slave/workspace/Geode- >>>> spark-connector/geode-assembly/build/install/apache- >>>> geode/lib/geode-dependencies.jar >>>> >>>> All WrappedArray(27897, 20289).length servers have been started >>>> Deploying:geode-functions_2.10-0.5.0.jar >>>> 16/09/23 16:12:09 WARN SparkContext: Another SparkContext is being >>>> constructed (or threw an exception in its constructor). This may >>>> indicate >>>> an error, since only one SparkContext may be running in this JVM (see >>>> SPARK-2243). The other SparkContext was created at: >>>> org.apache.spark.api.java.JavaSparkContext.<init>( >>>> JavaSparkContext.scala:61) >>>> ittest.org.apache.geode.spark.connector.JavaApiIntegrationTest. >>>> setUpBeforeClass(JavaApiIntegrationTest.java:75) >>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >>>> ssorImpl.java: >>>> 62) >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke( >>>> DelegatingMethodAccessorImpl.java:43) >>>> java.lang.reflect.Method.invoke(Method.java:497) >>>> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall( >>>> FrameworkMethod.java:47) >>>> org.junit.internal.runners.model.ReflectiveCallable.run( >>>> ReflectiveCallable.java:12) >>>> org.junit.runners.model.FrameworkMethod.invokeExplosively( >>>> FrameworkMethod.java:44) >>>> org.junit.internal.runners.statements.RunBefores. >>>> evaluate(RunBefores.java:24) >>>> org.junit.internal.runners.statements.RunAfters.evaluate( >>>> RunAfters.java:27) >>>> org.junit.runners.ParentRunner.run(ParentRunner.java:309) >>>> org.junit.runners.Suite.runChild(Suite.java:127) >>>> org.junit.runners.Suite.runChild(Suite.java:26) >>>> org.junit.runners.ParentRunner$3.run(ParentRunner.java:238) >>>> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63) >>>> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236) >>>> org.junit.runners.ParentRunner.access$000(ParentRunner.java:53) >>>> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229) >>>> org.junit.runners.ParentRunner.run(ParentRunner.java:309) >>>> [0m[ [0minfo [0m] [0m [31mException encountered when attempting to run >>>> a >>>> suite with class name: ittest.org.apache.geode.spark. >>>> connector.BasicIntegrationTest >>>> *** ABORTED *** [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m org.apache.spark.SparkException: Only one >>>> SparkContext may be running in this JVM (see SPARK-2243). To ignore this >>>> error, set spark.driver.allowMultipleContexts = true. The currently >>>> running SparkContext was created at: [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.apache.spark.SparkCont >>>> ext.<init>(SparkContext.scala:80) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >>>> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrat >>>> ionTest.scala:50) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.scalatest.BeforeAndAfterAll$class. >>>> beforeAll(BeforeAndAfterAll.scala:187) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >>>> RDDJoinRegionIntegrationTest.beforeAll(RDDJoinRegionIntegrat >>>> ionTest.scala:30) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.scalatest.BeforeAndAft >>>> erAll$class.run(BeforeAndAfterAll.scala:253) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mittest.org.apache.geode.spark.connector. >>>> RDDJoinRegionIntegrationTest.run(RDDJoinRegionIntegrationTest.scala:30) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.scalatest.tools.Framework.org >>>> $scalatest$tools$Framework$$runSuite(Framework.scala:462) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31morg.scalatest.tools.Framew >>>> ork$ScalaTestTask.execute(Framework.scala:671) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:294) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31msbt.ForkMain$Run$2.call(ForkMain.java:284) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >>>> FutureTask.run(FutureTask.java:266) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >>>> ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mjava.util.concurrent. >>>> ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31mjava.lang.Thread.run(Thread.java:745) [0m >>>> [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >>>> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(Spar >>>> kContext.scala:1811) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >>>> assertNoOtherContextIsRunning$1$$anonfun$apply$10.apply(Spar >>>> kContext.scala:1807) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at scala.Option.foreach(Option.scala:236) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >>>> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1807) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$$anonfun$ >>>> assertNoOtherContextIsRunning$1.apply(SparkContext.scala:1794) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at scala.Option.foreach(Option.scala:236) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$. >>>> assertNoOtherContextIsRunning(SparkContext.scala:1794) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext$ >>>> .markPartiallyConstructed(SparkContext.scala:1833) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at org.apache.spark.SparkContext. >>>> <init>(SparkContext.scala:89) >>>> [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m at ittest.org.apache.geode.spark. >>>> connector. >>>> BasicIntegrationTest.beforeAll(BasicIntegrationTest.scala:58) [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m ... [0m [0m >>>> [0m[ [0minfo [0m] [0mScalaTest [0m >>>> [0m[ [0minfo [0m] [0m [36mRun completed in 1 minute, 59 seconds. [0m >>>> [0m >>>> [0m[ [0minfo [0m] [0m [36mTotal number of tests run: 0 [0m [0m >>>> [0m[ [0minfo [0m] [0m [36mSuites: completed 1, aborted 3 [0m [0m >>>> [0m[ [0minfo [0m] [0m [36mTests: succeeded 0, failed 0, canceled 0, >>>> ignored 0, pending 0 [0m [0m >>>> [0m[ [0minfo [0m] [0m [31m*** 3 SUITES ABORTED *** [0m [0m >>>> [0m[ [31merror [0m] [0mError: Total 3, Failed 0, Errors 3, Passed 0 [0m >>>> [0m[ [31merror [0m] [0mError during tests: [0m >>>> [0m[ [31merror [0m] [0m ittest.org.apache.geode. >>>> spark.connector.RDDJoinRegionIntegrationTest >>>> [0m >>>> [0m[ [31merror [0m] [0m ittest.org.apache.geode. >>>> spark.connector.RetrieveRegionIntegrationTest >>>> [0m >>>> [0m[ [31merror [0m] [0m ittest.org.apache.geode.spark.connector. >>>> BasicIntegrationTest >>>> [0m >>>> [0m[ [31merror [0m] [0m(geode-spark-connector/it: [31mtest [0m) >>>> sbt.TestsFailedException: Tests unsuccessful [0m >>>> [0m[ [31merror [0m] [0mTotal time: 128 s, completed Sep 23, 2016 >>>> 4:12:09 >>>> PM [0m >>>> Build step 'Execute shell' marked build as failure >>>> Recording test results >>>> Skipped archiving because build is not successful >>>> >>>> >