[
https://issues.apache.org/jira/browse/GEODE-1560?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Anthony Baker closed GEODE-1560.
--------------------------------
> CI Failure: LocatorDUnitTest.testMultipleLocatorsRestartingAtSameTime NPE
> -------------------------------------------------------------------------
>
> Key: GEODE-1560
> URL: https://issues.apache.org/jira/browse/GEODE-1560
> Project: Geode
> Issue Type: Bug
> Components: membership
> Reporter: Scott Jewell
> Assignee: Hitesh Khamesra
> Labels: CI
> Fix For: 1.0.0-incubating.M3
>
>
> It appears to be perhaps a tainted environment or multiple tests running
> concurrently interfering with each other? Bottom of log shows this test
> experiencing multiple failures.
> Error Message
> java.lang.NullPointerException
> Stacktrace
> java.lang.NullPointerException
> at
> com.gemstone.gemfire.distributed.LocatorDUnitTest.testMultipleLocatorsRestartingAtSameTime(LocatorDUnitTest.java:1579)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> at
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> at
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> at
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> at
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> at
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> at org.junit.rules.RunRules.evaluate(RunRules.java:20)
> at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> at
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> at
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> at
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> at
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecuter.runTestClass(JUnitTestClassExecuter.java:112)
> at
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecuter.execute(JUnitTestClassExecuter.java:56)
> at
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassProcessor.processTestClass(JUnitTestClassProcessor.java:66)
> at
> org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:35)
> at
> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
> at
> org.gradle.messaging.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:32)
> at
> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:93)
> at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
> at
> org.gradle.api.internal.tasks.testing.worker.TestWorker.processTestClass(TestWorker.java:109)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:35)
> at
> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
> at
> org.gradle.messaging.remote.internal.hub.MessageHub$Handler.run(MessageHub.java:360)
> at
> org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:54)
> at
> org.gradle.internal.concurrent.StoppableExecutorImpl$1.run(StoppableExecutorImpl.java:40)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> Standard Output
> Executing
> [/home/jenkins/jenkins-slave/tools/hudson.model.JDK/latest1.8/jre/bin/java,
> -classpath,
> /home/jenkins/jenkins-slave/workspace/Geode-nightly/geode-core/build/classes/test:/home/jenkins/jenkins-slave/workspace/Geode-nightly/geode-core/build/resources/test:/home/jenkins/jenkins-slave/workspace/Geode-nightly/geode-core/build/classes/main:/home/jenkins/jenkins-slave/workspace/Geode-nightly/geode-core/build/resources/main:/home/jenkins/jenkins-slave/workspace/Geode-nightly/
> <snipped classpath>
> ...[truncated 252355 chars]...
> nordered connection
> [vm_0][info 2016/06/17 09:20:14.221 UTC <P2P message reader for
> 67.195.81.149(2109)<v5>:1030 shared unordered uid=2 port=60365> tid=0x4c]
> Performing final check for suspect member 67.195.81.149(2109)<v5>:1030
> reason=member unexpectedly shut down shared, unordered connection
> [vm_3][info 2016/06/17 09:20:14.223 UTC <P2P message reader for
> 67.195.81.149(2109)<v5>:1030 shared unordered uid=4 port=55419> tid=0x43]
> Performing final check for suspect member 67.195.81.149(2109)<v5>:1030
> reason=member unexpectedly shut down shared, unordered connection
> [info 2016/06/17 09:20:14.226 UTC <Geode Failure Detection Server thread 0>
> tid=0x53] GMSHealthMonitor server thread exiting
> [vm_2][info 2016/06/17 09:20:14.228 UTC <P2P message reader for
> 67.195.81.149(2109)<v5>:1030 shared unordered uid=5 port=47101> tid=0x4c]
> Performing final check for suspect member 67.195.81.149(2109)<v5>:1030
> reason=member unexpectedly shut down shared, unordered connection
> [vm_3][info 2016/06/17 09:20:14.228 UTC <P2P message reader for
> 67.195.81.149(2109)<v5>:1030 shared unordered uid=4 port=55419> tid=0x43]
> Final check failed - requesting removal of suspect member
> 67.195.81.149(2109)<v5>:1030
> [vm_1][info 2016/06/17 09:20:14.228 UTC <P2P message reader for
> 67.195.81.149(2109)<v5>:1030 shared unordered uid=1 port=41689> tid=0x4a]
> Final check failed - requesting removal of suspect member
> 67.195.81.149(2109)<v5>:1030
> [vm_4][info 2016/06/17 09:20:14.228 UTC <P2P message reader for
> 67.195.81.149(2109)<v5>:1030 shared unordered uid=3 port=49179> tid=0x42]
> Final check failed - requesting removal of suspect member
> 67.195.81.149(2109)<v5>:1030
> [vm_4][info 2016/06/17 09:20:14.230 UTC <unicast receiver,asf905-46898>
> tid=0x29] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2213)<v3>:1028 reason=member
> unexpectedly shut down shared, unordered connection
> [vm_2][info 2016/06/17 09:20:14.230 UTC <unicast receiver,asf905-29718>
> tid=0x2a] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2483)<v4>:1029 reason=member
> unexpectedly shut down shared, unordered connection
> [vm_0][info 2016/06/17 09:20:14.230 UTC <unicast receiver,asf905-22535>
> tid=0x2b] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2213)<v3>:1028 reason=member
> unexpectedly shut down shared, unordered connection
> [vm_2][info 2016/06/17 09:20:14.232 UTC <unicast receiver,asf905-29718>
> tid=0x2a] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2213)<v3>:1028 reason=member
> unexpectedly shut down shared, unordered connection
> [vm_3][info 2016/06/17 09:20:14.231 UTC <unicast receiver,asf905-50992>
> tid=0x29] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2191:locator)<ec><v1>:1026
> reason=member unexpectedly shut down shared, unordered connection
> [vm_3][info 2016/06/17 09:20:14.232 UTC <unicast receiver,asf905-50992>
> tid=0x29] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2483)<v4>:1029 reason=member
> unexpectedly shut down shared, unordered connection
> [vm_4][info 2016/06/17 09:20:14.232 UTC <unicast receiver,asf905-46898>
> tid=0x29] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2191:locator)<ec><v1>:1026
> reason=member unexpectedly shut down shared, unordered connection
> [vm_2][info 2016/06/17 09:20:14.232 UTC <P2P message reader for
> 67.195.81.149(2109)<v5>:1030 shared unordered uid=5 port=47101> tid=0x4c]
> Final check failed - requesting removal of suspect member
> 67.195.81.149(2109)<v5>:1030
> [vm_0][info 2016/06/17 09:20:14.231 UTC <unicast receiver,asf905-22535>
> tid=0x2b] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2483)<v4>:1029 reason=member
> unexpectedly shut down shared, unordered connection
> [vm_0][info 2016/06/17 09:20:14.233 UTC <unicast receiver,asf905-22535>
> tid=0x2b] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2191:locator)<ec><v1>:1026
> reason=member unexpectedly shut down shared, unordered connection
> [vm_2][info 2016/06/17 09:20:14.233 UTC <unicast receiver,asf905-29718>
> tid=0x2a] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2191:locator)<ec><v1>:1026
> reason=member unexpectedly shut down shared, unordered connection
> [vm_1][info 2016/06/17 09:20:14.230 UTC <unicast receiver,asf905-1766>
> tid=0x2a] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2213)<v3>:1028 reason=member
> unexpectedly shut down shared, unordered connection
> [vm_1][info 2016/06/17 09:20:14.231 UTC <unicast receiver,asf905-1766>
> tid=0x2a] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2483)<v4>:1029 reason=member
> unexpectedly shut down shared, unordered connection
> [info 2016/06/17 09:20:14.229 UTC <Geode Failure Detection Server thread 0>
> tid=0x53] GMSHealthMonitor server socket closed.
> [vm_0][info 2016/06/17 09:20:14.234 UTC <P2P message reader for
> 67.195.81.149(2109)<v5>:1030 shared unordered uid=2 port=60365> tid=0x4c]
> Final check passed for suspect member 67.195.81.149(2109)<v5>:1030
> [vm_1][info 2016/06/17 09:20:14.235 UTC <unicast receiver,asf905-1766>
> tid=0x2a] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2209:locator)<ec><v2>:1027
> reason=member unexpectedly shut down shared, unordered connection
> [info 2016/06/17 09:20:14.236 UTC <Test worker> tid=0x13] GMSHealthMonitor
> serverSocketExecutor is terminated
> [vm_0][info 2016/06/17 09:20:14.236 UTC <unicast receiver,asf905-22535>
> tid=0x2b] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2209:locator)<ec><v2>:1027
> reason=member unexpectedly shut down shared, unordered connection
> [vm_3][info 2016/06/17 09:20:14.236 UTC <unicast receiver,asf905-50992>
> tid=0x29] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2209:locator)<ec><v2>:1027
> reason=member unexpectedly shut down shared, unordered connection
> [vm_4][info 2016/06/17 09:20:14.236 UTC <unicast receiver,asf905-46898>
> tid=0x29] Membership received a request to remove
> 67.195.81.149(2109)<v5>:1030 from 67.195.81.149(2209:locator)<ec><v2>:1027
> reason=member unexpectedly shut down shared, unordered connection
> [vm_0][info 2016/06/17 09:20:14.247 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/605129648.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/605129648@3c685d9c
> [vm_0][info 2016/06/17 09:20:14.248 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_0] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/605129648.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/605129648@3c685d9c
> (took 0 ms)
> [vm_1][info 2016/06/17 09:20:14.251 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/770909585.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/770909585@59dc200e
> [vm_1][info 2016/06/17 09:20:14.252 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_1] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/770909585.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/770909585@59dc200e
> (took 0 ms)
> [vm_2][info 2016/06/17 09:20:14.254 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/571500128.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/571500128@74e866aa
> [vm_2][info 2016/06/17 09:20:14.254 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_2] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/571500128.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/571500128@74e866aa
> (took 0 ms)
> [vm_3][info 2016/06/17 09:20:14.257 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/1955872437.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/1955872437@7403f00f
> [vm_3][info 2016/06/17 09:20:14.258 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_3] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/1955872437.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$8/1955872437@7403f00f
> (took 0 ms)
> [vm_4][info 2016/06/17 09:20:14.269 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$6/259561976.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$6/259561976@ad5f76
> [vm_4][info 2016/06/17 09:20:14.270 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_4] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$6/259561976.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$6/259561976@ad5f76
> (took 0 ms)
> [info 2016/06/17 09:20:14.345 UTC <Test worker> tid=0x13] <ExpectedException
> action=remove>Removing shunned member</ExpectedException>
> [info 2016/06/17 09:20:14.346 UTC <Test worker> tid=0x13] <ExpectedException
> action=remove>Removing shunned member</ExpectedException>
> [vm_0][info 2016/06/17 09:20:14.352 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0 args on object:
> "IgnoredException remove"
> [vm_0][info 2016/06/17 09:20:14.353 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_0][info 2016/06/17 09:20:14.353 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_0][info 2016/06/17 09:20:14.353 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_0][info 2016/06/17 09:20:14.353 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_0] from com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0
> args on object: "IgnoredException remove" (took 0 ms)
> [vm_1][info 2016/06/17 09:20:14.359 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0 args on object:
> "IgnoredException remove"
> [vm_1][info 2016/06/17 09:20:14.359 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_1][info 2016/06/17 09:20:14.359 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_1][info 2016/06/17 09:20:14.359 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_1][info 2016/06/17 09:20:14.360 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_1] from com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0
> args on object: "IgnoredException remove" (took 0 ms)
> [vm_2][info 2016/06/17 09:20:14.364 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0 args on object:
> "IgnoredException remove"
> [vm_2][info 2016/06/17 09:20:14.364 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_2][info 2016/06/17 09:20:14.365 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_2][info 2016/06/17 09:20:14.365 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_2][info 2016/06/17 09:20:14.365 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_2] from com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0
> args on object: "IgnoredException remove" (took 0 ms)
> [vm_3][info 2016/06/17 09:20:14.370 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0 args on object:
> "IgnoredException remove"
> [vm_3][info 2016/06/17 09:20:14.370 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_3][info 2016/06/17 09:20:14.370 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_3][info 2016/06/17 09:20:14.370 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_3][info 2016/06/17 09:20:14.370 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_3] from com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0
> args on object: "IgnoredException remove" (took 0 ms)
> [vm_4][info 2016/06/17 09:20:14.380 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0 args on object:
> "IgnoredException remove"
> [vm_4][info 2016/06/17 09:20:14.381 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_4][info 2016/06/17 09:20:14.382 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_4][info 2016/06/17 09:20:14.382 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] <ExpectedException action=remove>Removing shunned
> member</ExpectedException>
> [vm_4][info 2016/06/17 09:20:14.383 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_4] from com.gemstone.gemfire.test.dunit.IgnoredException$1.run with 0
> args on object: "IgnoredException remove" (took 2 ms)
> [vm_0][info 2016/06/17 09:20:14.387 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/2045631151.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/2045631151@6673d4a1
> [vm_0][info 2016/06/17 09:20:14.419 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GemFireCache[id = 1505642518; isClosing = true; isShutDownAll =
> false; created = Fri Jun 17 09:19:49 UTC 2016; server = false; copyOnRead =
> false; lockLease = 120; lockTimeout = 60]: Now closing.
> [vm_0][info 2016/06/17 09:20:14.469 UTC
> <com.gemstone.gemfire.distributed.internal.InternalLocator$1@1a102134>
> tid=0x4f] Stopping Distribution Locator on asf905.gq1.ygridcore.net[22591]
> [vm_0][info 2016/06/17 09:20:14.484 UTC <Distribution Locator on
> asf905.gq1.ygridcore.net[null]> tid=0x1d] locator shutting down
> [vm_0][info 2016/06/17 09:20:14.487 UTC <Distribution Locator on
> asf905.gq1.ygridcore.net[null]> tid=0x1d] Closing locator's cache
> [vm_0][info 2016/06/17 09:20:14.487 UTC
> <com.gemstone.gemfire.distributed.internal.InternalLocator$1@1a102134>
> tid=0x4f] Distribution Locator on asf905.gq1.ygridcore.net[0] is stopped
> [vm_0][info 2016/06/17 09:20:14.490 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Shutting down DistributionManager
> 67.195.81.149(2186:locator)<ec><v0>:1025.
> [vm_3][info 2016/06/17 09:20:14.493 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] received leave request from
> 67.195.81.149(2186:locator)<ec><v0>:1025 for
> 67.195.81.149(2186:locator)<ec><v0>:1025
> [vm_2][info 2016/06/17 09:20:14.493 UTC <Pooled High Priority Message
> Processor 2> tid=0x43] received leave request from
> 67.195.81.149(2186:locator)<ec><v0>:1025 for
> 67.195.81.149(2186:locator)<ec><v0>:1025
> [vm_1][info 2016/06/17 09:20:14.493 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] received leave request from
> 67.195.81.149(2186:locator)<ec><v0>:1025 for
> 67.195.81.149(2186:locator)<ec><v0>:1025
> [vm_1][info 2016/06/17 09:20:14.494 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] This member is becoming the membership coordinator
> with address 67.195.81.149(2191:locator)<ec><v1>:1026
> [vm_3][info 2016/06/17 09:20:14.494 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] Member at 67.195.81.149(2186:locator)<ec><v0>:1025
> gracefully left the distributed cache: shutdown message received
> [vm_4][info 2016/06/17 09:20:14.494 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] received leave request from
> 67.195.81.149(2186:locator)<ec><v0>:1025 for
> 67.195.81.149(2186:locator)<ec><v0>:1025
> [vm_2][info 2016/06/17 09:20:14.495 UTC <Pooled High Priority Message
> Processor 2> tid=0x43] Member at 67.195.81.149(2186:locator)<ec><v0>:1025
> gracefully left the distributed cache: shutdown message received
> [vm_4][info 2016/06/17 09:20:14.496 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] Member at 67.195.81.149(2186:locator)<ec><v0>:1025
> gracefully left the distributed cache: shutdown message received
> [vm_1][info 2016/06/17 09:20:14.498 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] ViewCreator starting
> on:67.195.81.149(2191:locator)<ec><v1>:1026
> [vm_1][info 2016/06/17 09:20:14.499 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] Member at 67.195.81.149(2186:locator)<ec><v0>:1025
> gracefully left the distributed cache: shutdown message received
> [vm_1][info 2016/06/17 09:20:14.499 UTC <Geode Membership View Creator>
> tid=0x4d] View Creator thread is starting
> [vm_1][info 2016/06/17 09:20:14.500 UTC <Geode Membership View Creator>
> tid=0x4d] 67.195.81.149(2109)<v5>:1030 had a weight of 10
> [vm_1][info 2016/06/17 09:20:14.501 UTC <Geode Membership View Creator>
> tid=0x4d] preparing new view
> View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_1]failure detection ports: 18067 4457 8640 12514
> [vm_1][info 2016/06/17 09:20:14.505 UTC <Geode Membership View Creator>
> tid=0x4d] finished waiting for responses to view preparation
> [vm_1][info 2016/06/17 09:20:14.506 UTC <Geode Membership View Creator>
> tid=0x4d] received new view:
> View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_1]old view is: View[67.195.81.149(2186:locator)<ec><v0>:1025|5] members:
> [67.195.81.149(2186:locator)<ec><v0>:1025,
> 67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029, 67.195.81.149(2109)<v5>:1030]
> [vm_1][info 2016/06/17 09:20:14.506 UTC <Geode Membership View Creator>
> tid=0x4d] 67.195.81.149(2109)<v5>:1030 had a weight of 10
> [vm_1][info 2016/06/17 09:20:14.506 UTC <Geode Membership View Creator>
> tid=0x4d] Peer locator received new membership view:
> View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_1][info 2016/06/17 09:20:14.508 UTC <Geode Membership View Creator>
> tid=0x4d] sending new view View[67.195.81.149(2191:locator)<ec><v1>:1026|19]
> members: [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_1]failure detection ports: 18067 4457 8640 12514
> [vm_1][info 2016/06/17 09:20:14.510 UTC <View Message Processor> tid=0x42] I,
> 67.195.81.149(2191:locator)<ec><v1>:1026, am the elder.
> [vm_4][info 2016/06/17 09:20:14.511 UTC <unicast receiver,asf905-46898>
> tid=0x29] received new view:
> View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_4]old view is: View[67.195.81.149(2186:locator)<ec><v0>:1025|5] members:
> [67.195.81.149(2186:locator)<ec><v0>:1025,
> 67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029, 67.195.81.149(2109)<v5>:1030]
> [vm_3][info 2016/06/17 09:20:14.511 UTC <unicast receiver,asf905-50992>
> tid=0x29] received new view:
> View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_3]old view is: View[67.195.81.149(2186:locator)<ec><v0>:1025|5] members:
> [67.195.81.149(2186:locator)<ec><v0>:1025,
> 67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029, 67.195.81.149(2109)<v5>:1030]
> [vm_2][info 2016/06/17 09:20:14.511 UTC <unicast receiver,asf905-29718>
> tid=0x2a] received new view:
> View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_2]old view is: View[67.195.81.149(2186:locator)<ec><v0>:1025|5] members:
> [67.195.81.149(2186:locator)<ec><v0>:1025,
> 67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029, 67.195.81.149(2109)<v5>:1030]
> [vm_2][info 2016/06/17 09:20:14.511 UTC <unicast receiver,asf905-29718>
> tid=0x2a] Peer locator received new membership view:
> View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_0][info 2016/06/17 09:20:14.531 UTC <Geode Membership View Creator>
> tid=0x31] View Creator is processing 1 requests for the next membership view
> [vm_0][info 2016/06/17 09:20:14.532 UTC <Geode Membership View Creator>
> tid=0x31] 67.195.81.149(2109)<v5>:1030 had a weight of 10
> [vm_0][info 2016/06/17 09:20:14.533 UTC <Geode Membership View Creator>
> tid=0x31] preparing new view View[67.195.81.149(2186:locator)<ec><v0>:1025|6]
> members: [67.195.81.149(2186:locator)<ec><v0>:1025,
> 67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] crashed: [67.195.81.149(2109)<v5>:1030]
> [vm_0]failure detection ports: 57887 18067 4457 8640 12514
> [vm_3][info 2016/06/17 09:20:14.534 UTC <unicast receiver,asf905-50992>
> tid=0x29] Ignoring the view View[67.195.81.149(2186:locator)<ec><v0>:1025|6]
> members: [67.195.81.149(2186:locator)<ec><v0>:1025,
> 67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] crashed: [67.195.81.149(2109)<v5>:1030] from
> member 67.195.81.149(2186:locator)<ec><v0>:1025, which is not in my current
> view View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_1][info 2016/06/17 09:20:14.534 UTC <unicast receiver,asf905-1766>
> tid=0x2a] Ignoring the view View[67.195.81.149(2186:locator)<ec><v0>:1025|6]
> members: [67.195.81.149(2186:locator)<ec><v0>:1025,
> 67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] crashed: [67.195.81.149(2109)<v5>:1030] from
> member 67.195.81.149(2186:locator)<ec><v0>:1025, which is not in my current
> view View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_2][info 2016/06/17 09:20:14.535 UTC <unicast receiver,asf905-29718>
> tid=0x2a] Ignoring the view View[67.195.81.149(2186:locator)<ec><v0>:1025|6]
> members: [67.195.81.149(2186:locator)<ec><v0>:1025,
> 67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] crashed: [67.195.81.149(2109)<v5>:1030] from
> member 67.195.81.149(2186:locator)<ec><v0>:1025, which is not in my current
> view View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_4][info 2016/06/17 09:20:14.535 UTC <unicast receiver,asf905-46898>
> tid=0x29] Ignoring the view View[67.195.81.149(2186:locator)<ec><v0>:1025|6]
> members: [67.195.81.149(2186:locator)<ec><v0>:1025,
> 67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] crashed: [67.195.81.149(2109)<v5>:1030] from
> member 67.195.81.149(2186:locator)<ec><v0>:1025, which is not in my current
> view View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_0][info 2016/06/17 09:20:14.598 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Now closing distribution for
> 67.195.81.149(2186:locator)<ec><v0>:1025
> [vm_0][info 2016/06/17 09:20:14.599 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Stopping membership services
> [vm_2][info 2016/06/17 09:20:14.601 UTC <unicast receiver,asf905-29718>
> tid=0x2a] received leave request from
> 67.195.81.149(2186:locator)<ec><v0>:1025 for
> 67.195.81.149(2186:locator)<ec><v0>:1025
> [vm_4][info 2016/06/17 09:20:14.601 UTC <unicast receiver,asf905-46898>
> tid=0x29] received leave request from
> 67.195.81.149(2186:locator)<ec><v0>:1025 for
> 67.195.81.149(2186:locator)<ec><v0>:1025
> [vm_0][info 2016/06/17 09:20:14.601 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor server socket is closed in stopServices().
> [vm_1][info 2016/06/17 09:20:14.601 UTC <unicast receiver,asf905-1766>
> tid=0x2a] received leave request from
> 67.195.81.149(2186:locator)<ec><v0>:1025 for
> 67.195.81.149(2186:locator)<ec><v0>:1025
> [vm_0][info 2016/06/17 09:20:14.601 UTC <Geode Failure Detection Server
> thread 0> tid=0x2f] GMSHealthMonitor server thread exiting
> [vm_3][info 2016/06/17 09:20:14.601 UTC <unicast receiver,asf905-50992>
> tid=0x29] received leave request from
> 67.195.81.149(2186:locator)<ec><v0>:1025 for
> 67.195.81.149(2186:locator)<ec><v0>:1025
> [vm_0][info 2016/06/17 09:20:14.602 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor serverSocketExecutor is terminated
> [vm_0][info 2016/06/17 09:20:14.613 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] DistributionManager stopped in 123ms.
> [vm_0][info 2016/06/17 09:20:14.613 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Marking DistributionManager
> 67.195.81.149(2186:locator)<ec><v0>:1025 as closed.
> [vm_0][info 2016/06/17 09:20:14.665 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_0] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/2045631151.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/2045631151@6673d4a1
> (took 277 ms)
> [vm_1][info 2016/06/17 09:20:14.667 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/1842040959.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/1842040959@19e3269f
> [vm_1][info 2016/06/17 09:20:14.695 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GemFireCache[id = 489661411; isClosing = true; isShutDownAll =
> false; created = Fri Jun 17 09:19:51 UTC 2016; server = false; copyOnRead =
> false; lockLease = 120; lockTimeout = 60]: Now closing.
> [vm_1][info 2016/06/17 09:20:14.747 UTC
> <com.gemstone.gemfire.distributed.internal.InternalLocator$1@1b90f4e4>
> tid=0x51] Stopping Distribution Locator on asf905.gq1.ygridcore.net[23732]
> [vm_1][info 2016/06/17 09:20:14.750 UTC <Distribution Locator on
> asf905.gq1.ygridcore.net[null]> tid=0x1d] locator shutting down
> [vm_1][info 2016/06/17 09:20:14.752 UTC <Distribution Locator on
> asf905.gq1.ygridcore.net[null]> tid=0x1d] Closing locator's cache
> [vm_1][info 2016/06/17 09:20:14.752 UTC
> <com.gemstone.gemfire.distributed.internal.InternalLocator$1@1b90f4e4>
> tid=0x51] Distribution Locator on asf905.gq1.ygridcore.net[0] is stopped
> [vm_1][info 2016/06/17 09:20:14.754 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Shutting down DistributionManager
> 67.195.81.149(2191:locator)<ec><v1>:1026.
> [vm_3][info 2016/06/17 09:20:14.756 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] received leave request from
> 67.195.81.149(2191:locator)<ec><v1>:1026 for
> 67.195.81.149(2191:locator)<ec><v1>:1026
> [vm_2][info 2016/06/17 09:20:14.756 UTC <Pooled High Priority Message
> Processor 2> tid=0x43] received leave request from
> 67.195.81.149(2191:locator)<ec><v1>:1026 for
> 67.195.81.149(2191:locator)<ec><v1>:1026
> [vm_4][info 2016/06/17 09:20:14.756 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] received leave request from
> 67.195.81.149(2191:locator)<ec><v1>:1026 for
> 67.195.81.149(2191:locator)<ec><v1>:1026
> [vm_2][info 2016/06/17 09:20:14.757 UTC <Pooled High Priority Message
> Processor 2> tid=0x43] This member is becoming the membership coordinator
> with address 67.195.81.149(2209:locator)<ec><v2>:1027
> [vm_4][info 2016/06/17 09:20:14.757 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] Member at 67.195.81.149(2191:locator)<ec><v1>:1026
> gracefully left the distributed cache: shutdown message received
> [vm_3][info 2016/06/17 09:20:14.757 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] Member at 67.195.81.149(2191:locator)<ec><v1>:1026
> gracefully left the distributed cache: shutdown message received
> [vm_2][info 2016/06/17 09:20:14.760 UTC <Pooled High Priority Message
> Processor 2> tid=0x43] ViewCreator starting
> on:67.195.81.149(2209:locator)<ec><v2>:1027
> [vm_2][info 2016/06/17 09:20:14.760 UTC <Pooled High Priority Message
> Processor 2> tid=0x43] Member at 67.195.81.149(2191:locator)<ec><v1>:1026
> gracefully left the distributed cache: shutdown message received
> [vm_2][info 2016/06/17 09:20:14.761 UTC <Geode Membership View Creator>
> tid=0x52] View Creator thread is starting
> [vm_2][info 2016/06/17 09:20:14.762 UTC <Geode Membership View Creator>
> tid=0x52] preparing new view
> View[67.195.81.149(2209:locator)<ec><v2>:1027|26] members:
> [67.195.81.149(2209:locator)<ec><v2>:1027,
> 67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2191:locator)<ec><v1>:1026]
> [vm_2]failure detection ports: 4457 8640 12514
> [vm_2][info 2016/06/17 09:20:14.765 UTC <Geode Membership View Creator>
> tid=0x52] finished waiting for responses to view preparation
> [vm_2][info 2016/06/17 09:20:14.765 UTC <Geode Membership View Creator>
> tid=0x52] received new view:
> View[67.195.81.149(2209:locator)<ec><v2>:1027|26] members:
> [67.195.81.149(2209:locator)<ec><v2>:1027,
> 67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2191:locator)<ec><v1>:1026]
> [vm_2]old view is: View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_2][info 2016/06/17 09:20:14.765 UTC <Geode Membership View Creator>
> tid=0x52] Peer locator received new membership view:
> View[67.195.81.149(2209:locator)<ec><v2>:1027|26] members:
> [67.195.81.149(2209:locator)<ec><v2>:1027,
> 67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2191:locator)<ec><v1>:1026]
> [vm_2][info 2016/06/17 09:20:14.767 UTC <Geode Membership View Creator>
> tid=0x52] sending new view View[67.195.81.149(2209:locator)<ec><v2>:1027|26]
> members: [67.195.81.149(2209:locator)<ec><v2>:1027,
> 67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2191:locator)<ec><v1>:1026]
> [vm_2]failure detection ports: 4457 8640 12514
> [vm_2][info 2016/06/17 09:20:14.768 UTC <View Message Processor> tid=0x46] I,
> 67.195.81.149(2209:locator)<ec><v2>:1027, am the elder.
> [vm_3][info 2016/06/17 09:20:14.769 UTC <unicast receiver,asf905-50992>
> tid=0x29] received new view:
> View[67.195.81.149(2209:locator)<ec><v2>:1027|26] members:
> [67.195.81.149(2209:locator)<ec><v2>:1027,
> 67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2191:locator)<ec><v1>:1026]
> [vm_4][info 2016/06/17 09:20:14.769 UTC <unicast receiver,asf905-46898>
> tid=0x29] received new view:
> View[67.195.81.149(2209:locator)<ec><v2>:1027|26] members:
> [67.195.81.149(2209:locator)<ec><v2>:1027,
> 67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2191:locator)<ec><v1>:1026]
> [vm_4]old view is: View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_3]old view is: View[67.195.81.149(2191:locator)<ec><v1>:1026|19] members:
> [67.195.81.149(2191:locator)<ec><v1>:1026,
> 67.195.81.149(2209:locator)<ec><v2>:1027, 67.195.81.149(2213)<v3>:1028{lead},
> 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2186:locator)<ec><v0>:1025] crashed:
> [67.195.81.149(2109)<v5>:1030]
> [vm_1][info 2016/06/17 09:20:14.857 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Now closing distribution for
> 67.195.81.149(2191:locator)<ec><v1>:1026
> [vm_1][info 2016/06/17 09:20:14.858 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Stopping membership services
> [vm_4][info 2016/06/17 09:20:14.860 UTC <unicast receiver,asf905-46898>
> tid=0x29] received leave request from
> 67.195.81.149(2191:locator)<ec><v1>:1026 for
> 67.195.81.149(2191:locator)<ec><v1>:1026
> [vm_2][info 2016/06/17 09:20:14.860 UTC <unicast receiver,asf905-29718>
> tid=0x2a] received leave request from
> 67.195.81.149(2191:locator)<ec><v1>:1026 for
> 67.195.81.149(2191:locator)<ec><v1>:1026
> [vm_1][info 2016/06/17 09:20:14.860 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor server socket is closed in stopServices().
> [vm_1][info 2016/06/17 09:20:14.860 UTC <Geode Failure Detection Server
> thread 0> tid=0x2e] GMSHealthMonitor server thread exiting
> [vm_3][info 2016/06/17 09:20:14.860 UTC <unicast receiver,asf905-50992>
> tid=0x29] received leave request from
> 67.195.81.149(2191:locator)<ec><v1>:1026 for
> 67.195.81.149(2191:locator)<ec><v1>:1026
> [vm_1][info 2016/06/17 09:20:14.861 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor serverSocketExecutor is terminated
> [vm_1][info 2016/06/17 09:20:14.870 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] DistributionManager stopped in 116ms.
> [vm_1][info 2016/06/17 09:20:14.870 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Marking DistributionManager
> 67.195.81.149(2191:locator)<ec><v1>:1026 as closed.
> [vm_1][info 2016/06/17 09:20:14.920 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_1] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/1842040959.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/1842040959@19e3269f
> (took 252 ms)
> [vm_2][info 2016/06/17 09:20:14.923 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/191678626.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/191678626@6f827b31
> [vm_2][info 2016/06/17 09:20:14.945 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GemFireCache[id = 1961479738; isClosing = true; isShutDownAll =
> false; created = Fri Jun 17 09:19:54 UTC 2016; server = false; copyOnRead =
> false; lockLease = 120; lockTimeout = 60]: Now closing.
> [vm_2][info 2016/06/17 09:20:14.984 UTC
> <com.gemstone.gemfire.distributed.internal.InternalLocator$1@6e462170>
> tid=0x55] Stopping Distribution Locator on asf905.gq1.ygridcore.net[29658]
> [vm_2][info 2016/06/17 09:20:14.987 UTC <Distribution Locator on
> asf905.gq1.ygridcore.net[null]> tid=0x1d] locator shutting down
> [vm_2][info 2016/06/17 09:20:14.989 UTC <Distribution Locator on
> asf905.gq1.ygridcore.net[null]> tid=0x1d] Closing locator's cache
> [vm_2][info 2016/06/17 09:20:14.989 UTC
> <com.gemstone.gemfire.distributed.internal.InternalLocator$1@6e462170>
> tid=0x55] Distribution Locator on asf905.gq1.ygridcore.net[0] is stopped
> [vm_2][info 2016/06/17 09:20:14.991 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Shutting down DistributionManager
> 67.195.81.149(2209:locator)<ec><v2>:1027.
> [vm_4][info 2016/06/17 09:20:14.993 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] received leave request from
> 67.195.81.149(2209:locator)<ec><v2>:1027 for
> 67.195.81.149(2209:locator)<ec><v2>:1027
> [vm_3][info 2016/06/17 09:20:14.993 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] received leave request from
> 67.195.81.149(2209:locator)<ec><v2>:1027 for
> 67.195.81.149(2209:locator)<ec><v2>:1027
> [vm_3][info 2016/06/17 09:20:14.994 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] This member is becoming the membership coordinator
> with address 67.195.81.149(2213)<v3>:1028
> [vm_4][info 2016/06/17 09:20:14.994 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] Member at 67.195.81.149(2209:locator)<ec><v2>:1027
> gracefully left the distributed cache: shutdown message received
> [vm_3][info 2016/06/17 09:20:14.998 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] ViewCreator starting on:67.195.81.149(2213)<v3>:1028
> [vm_3][info 2016/06/17 09:20:14.998 UTC <Geode Membership View Creator>
> tid=0x4b] View Creator thread is starting
> [vm_3][info 2016/06/17 09:20:14.999 UTC <Pooled High Priority Message
> Processor 1> tid=0x3f] Member at 67.195.81.149(2209:locator)<ec><v2>:1027
> gracefully left the distributed cache: shutdown message received
> [vm_3][info 2016/06/17 09:20:15.000 UTC <Geode Membership View Creator>
> tid=0x4b] preparing new view View[67.195.81.149(2213)<v3>:1028|37] members:
> [67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2209:locator)<ec><v2>:1027]
> [vm_3]failure detection ports: 8640 12514
> [vm_3][info 2016/06/17 09:20:15.003 UTC <Geode Membership View Creator>
> tid=0x4b] finished waiting for responses to view preparation
> [vm_3][info 2016/06/17 09:20:15.004 UTC <Geode Membership View Creator>
> tid=0x4b] received new view: View[67.195.81.149(2213)<v3>:1028|37] members:
> [67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2209:locator)<ec><v2>:1027]
> [vm_3]old view is: View[67.195.81.149(2209:locator)<ec><v2>:1027|26] members:
> [67.195.81.149(2209:locator)<ec><v2>:1027,
> 67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2191:locator)<ec><v1>:1026]
> [vm_3][info 2016/06/17 09:20:15.005 UTC <Geode Membership View Creator>
> tid=0x4b] sending new view View[67.195.81.149(2213)<v3>:1028|37] members:
> [67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2209:locator)<ec><v2>:1027]
> [vm_3]failure detection ports: 8640 12514
> [vm_3][info 2016/06/17 09:20:15.006 UTC <View Message Processor> tid=0x3c] I,
> 67.195.81.149(2213)<v3>:1028, am the elder.
> [vm_4][info 2016/06/17 09:20:15.007 UTC <unicast receiver,asf905-46898>
> tid=0x29] received new view: View[67.195.81.149(2213)<v3>:1028|37] members:
> [67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2209:locator)<ec><v2>:1027]
> [vm_4]old view is: View[67.195.81.149(2209:locator)<ec><v2>:1027|26] members:
> [67.195.81.149(2209:locator)<ec><v2>:1027,
> 67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2191:locator)<ec><v1>:1026]
> [vm_2][info 2016/06/17 09:20:15.094 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Now closing distribution for
> 67.195.81.149(2209:locator)<ec><v2>:1027
> [vm_2][info 2016/06/17 09:20:15.094 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Stopping membership services
> [vm_3][info 2016/06/17 09:20:15.096 UTC <unicast receiver,asf905-50992>
> tid=0x29] received leave request from
> 67.195.81.149(2209:locator)<ec><v2>:1027 for
> 67.195.81.149(2209:locator)<ec><v2>:1027
> [vm_4][info 2016/06/17 09:20:15.096 UTC <unicast receiver,asf905-46898>
> tid=0x29] received leave request from
> 67.195.81.149(2209:locator)<ec><v2>:1027 for
> 67.195.81.149(2209:locator)<ec><v2>:1027
> [vm_2][info 2016/06/17 09:20:15.095 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor server socket is closed in stopServices().
> [vm_2][info 2016/06/17 09:20:15.096 UTC <Geode Failure Detection Server
> thread 0> tid=0x2e] GMSHealthMonitor server thread exiting
> [vm_2][info 2016/06/17 09:20:15.096 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor serverSocketExecutor is terminated
> [vm_2][info 2016/06/17 09:20:15.103 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] DistributionManager stopped in 112ms.
> [vm_2][info 2016/06/17 09:20:15.103 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Marking DistributionManager
> 67.195.81.149(2209:locator)<ec><v2>:1027 as closed.
> [vm_2][info 2016/06/17 09:20:15.151 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_2] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/191678626.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/191678626@6f827b31
> (took 228 ms)
> [vm_3][info 2016/06/17 09:20:15.154 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/465380541.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/465380541@35024eff
> [vm_3][info 2016/06/17 09:20:15.155 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Shutting down DistributionManager 67.195.81.149(2213)<v3>:1028.
> [vm_4][info 2016/06/17 09:20:15.158 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] received leave request from
> 67.195.81.149(2213)<v3>:1028 for 67.195.81.149(2213)<v3>:1028
> [vm_4][info 2016/06/17 09:20:15.159 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] This member is becoming the membership coordinator
> with address 67.195.81.149(2483)<v4>:1029
> [vm_4][info 2016/06/17 09:20:15.163 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] ViewCreator starting on:67.195.81.149(2483)<v4>:1029
> [vm_4][info 2016/06/17 09:20:15.163 UTC <Geode Membership View Creator>
> tid=0x4e] View Creator thread is starting
> [vm_4][info 2016/06/17 09:20:15.164 UTC <Geode Membership View Creator>
> tid=0x4e] received new view: View[67.195.81.149(2483)<v4>:1029|49] members:
> [67.195.81.149(2483)<v4>:1029{lead}] shutdown: [67.195.81.149(2213)<v3>:1028]
> [vm_4]old view is: View[67.195.81.149(2213)<v3>:1028|37] members:
> [67.195.81.149(2213)<v3>:1028{lead}, 67.195.81.149(2483)<v4>:1029] shutdown:
> [67.195.81.149(2209:locator)<ec><v2>:1027]
> [vm_4][info 2016/06/17 09:20:15.164 UTC <Pooled High Priority Message
> Processor 1> tid=0x43] Member at 67.195.81.149(2213)<v3>:1028 gracefully left
> the distributed cache: shutdown message received
> [vm_4][info 2016/06/17 09:20:15.165 UTC <Geode Membership View Creator>
> tid=0x4e] no recipients for new view aside from myself
> [vm_4][info 2016/06/17 09:20:15.166 UTC <View Message Processor> tid=0x40]
> 67.195.81.149(2483)<v4>:1029 is the elder and the only member.
> [vm_3][info 2016/06/17 09:20:15.261 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Now closing distribution for 67.195.81.149(2213)<v3>:1028
> [vm_3][info 2016/06/17 09:20:15.261 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Stopping membership services
> [vm_3][info 2016/06/17 09:20:15.263 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor server socket is closed in stopServices().
> [vm_4][info 2016/06/17 09:20:15.263 UTC <unicast receiver,asf905-46898>
> tid=0x29] received leave request from 67.195.81.149(2213)<v3>:1028 for
> 67.195.81.149(2213)<v3>:1028
> [vm_3][info 2016/06/17 09:20:15.263 UTC <Geode Failure Detection Server
> thread 0> tid=0x2d] GMSHealthMonitor server thread exiting
> [vm_3][info 2016/06/17 09:20:15.264 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor serverSocketExecutor is terminated
> [vm_3][info 2016/06/17 09:20:15.273 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] DistributionManager stopped in 117ms.
> [vm_3][info 2016/06/17 09:20:15.274 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Marking DistributionManager 67.195.81.149(2213)<v3>:1028 as closed.
> [vm_3][info 2016/06/17 09:20:15.366 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_3] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/465380541.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$9/465380541@35024eff
> (took 211 ms)
> [vm_4][info 2016/06/17 09:20:15.373 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$7/1904510272.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$7/1904510272@371b8f87
> [vm_4][info 2016/06/17 09:20:15.375 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Shutting down DistributionManager 67.195.81.149(2483)<v4>:1029.
> [vm_4][info 2016/06/17 09:20:15.481 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Now closing distribution for 67.195.81.149(2483)<v4>:1029
> [vm_4][info 2016/06/17 09:20:15.481 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Stopping membership services
> [vm_4][info 2016/06/17 09:20:15.483 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor server socket is closed in stopServices().
> [vm_4][info 2016/06/17 09:20:15.484 UTC <Geode Failure Detection Server
> thread 0> tid=0x2d] GMSHealthMonitor server thread exiting
> [vm_4][info 2016/06/17 09:20:15.484 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] GMSHealthMonitor serverSocketExecutor is terminated
> [vm_4][info 2016/06/17 09:20:15.493 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] DistributionManager stopped in 118ms.
> [vm_4][info 2016/06/17 09:20:15.493 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Marking DistributionManager 67.195.81.149(2483)<v4>:1029 as closed.
> [vm_4][info 2016/06/17 09:20:15.590 UTC <RMI TCP Connection(1)-67.195.81.149>
> tid=0x1b] Got result: null
> [vm_4] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$7/1904510272.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$7/1904510272@371b8f87
> (took 216 ms)
> [locator][info 2016/06/17 09:20:15.611 UTC <RMI TCP
> Connection(2)-67.195.81.149> tid=0x1b] Received method:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$5/28871561.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$5/28871561@60944f82
> [locator][info 2016/06/17 09:20:15.616 UTC <RMI TCP
> Connection(2)-67.195.81.149> tid=0x1b] Got result: null
> [locator] from
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$5/28871561.run
> with 0 args on object:
> com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$5/28871561@60944f82
> (took 4 ms)
> Standard Error
> Suspicious strings were written to the log during this run.
> Fix the strings or use IgnoredException.addIgnoredException to ignore.
> -----------------------------------------------------------------------
> Found suspect string in log4j at line 2738
> [error 2016/06/17 09:20:00.793 UTC <Geode UDP INT-1,asf905-22535> tid=0x42]
> Exception deserializing message payload: [dst: 67.195.81.149<v0>:1025, src:
> 67.195.81.149<v5>:1030 (2 headers), size=189 bytes,
> flags=OOB|NO_RELIABILITY|INTERNAL]
> java.io.IOException: Unknown header byte: 0
> at
> com.gemstone.gemfire.internal.InternalDataSerializer.basicReadObject(InternalDataSerializer.java:2979)
> at
> com.gemstone.gemfire.DataSerializer.readObject(DataSerializer.java:3271)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger.readJGMessage(JGroupsMessenger.java:849)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1043)
> at org.jgroups.JChannel.invokeCallback(JChannel.java:816)
> at org.jgroups.JChannel.up(JChannel.java:741)
> at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030)
> at org.jgroups.protocols.FRAG2.up(FRAG2.java:165)
> at org.jgroups.protocols.FlowControl.up(FlowControl.java:381)
> at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:435)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75)
> at org.jgroups.protocols.TP.passMessageUp(TP.java:1567)
> at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1783)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> -----------------------------------------------------------------------
> Found suspect string in log4j at line 2758
> [error 2016/06/17 09:20:03.500 UTC <Geode UDP INT-2,asf905-22535> tid=0x45]
> Exception deserializing message payload: [dst: 67.195.81.149<v0>:1025, src:
> 67.195.81.149<v5>:1030 (2 headers), size=189 bytes,
> flags=OOB|NO_RELIABILITY|INTERNAL]
> java.io.IOException: Unknown header byte: 0
> at
> com.gemstone.gemfire.internal.InternalDataSerializer.basicReadObject(InternalDataSerializer.java:2979)
> at
> com.gemstone.gemfire.DataSerializer.readObject(DataSerializer.java:3271)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger.readJGMessage(JGroupsMessenger.java:849)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1043)
> at org.jgroups.JChannel.invokeCallback(JChannel.java:816)
> at org.jgroups.JChannel.up(JChannel.java:741)
> at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030)
> at org.jgroups.protocols.FRAG2.up(FRAG2.java:165)
> at org.jgroups.protocols.FlowControl.up(FlowControl.java:381)
> at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:435)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75)
> at org.jgroups.protocols.TP.passMessageUp(TP.java:1567)
> at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1783)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> -----------------------------------------------------------------------
> Found suspect string in log4j at line 2778
> [error 2016/06/17 09:20:06.824 UTC <Geode UDP INT-1,asf905-22535> tid=0x42]
> Exception deserializing message payload: [dst: 67.195.81.149<v0>:1025, src:
> 67.195.81.149<v5>:1030 (2 headers), size=189 bytes,
> flags=OOB|NO_RELIABILITY|INTERNAL]
> java.io.IOException: Unknown header byte: 0
> at
> com.gemstone.gemfire.internal.InternalDataSerializer.basicReadObject(InternalDataSerializer.java:2979)
> at
> com.gemstone.gemfire.DataSerializer.readObject(DataSerializer.java:3271)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger.readJGMessage(JGroupsMessenger.java:849)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1043)
> at org.jgroups.JChannel.invokeCallback(JChannel.java:816)
> at org.jgroups.JChannel.up(JChannel.java:741)
> at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030)
> at org.jgroups.protocols.FRAG2.up(FRAG2.java:165)
> at org.jgroups.protocols.FlowControl.up(FlowControl.java:381)
> at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:435)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75)
> at org.jgroups.protocols.TP.passMessageUp(TP.java:1567)
> at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1783)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> -----------------------------------------------------------------------
> Found suspect string in log4j at line 2798
> [error 2016/06/17 09:20:09.835 UTC <Geode UDP INT-2,asf905-22535> tid=0x45]
> Exception deserializing message payload: [dst: 67.195.81.149<v0>:1025, src:
> 67.195.81.149<v5>:1030 (2 headers), size=189 bytes,
> flags=OOB|NO_RELIABILITY|INTERNAL]
> java.io.IOException: Unknown header byte: 0
> at
> com.gemstone.gemfire.internal.InternalDataSerializer.basicReadObject(InternalDataSerializer.java:2979)
> at
> com.gemstone.gemfire.DataSerializer.readObject(DataSerializer.java:3271)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger.readJGMessage(JGroupsMessenger.java:849)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1043)
> at org.jgroups.JChannel.invokeCallback(JChannel.java:816)
> at org.jgroups.JChannel.up(JChannel.java:741)
> at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030)
> at org.jgroups.protocols.FRAG2.up(FRAG2.java:165)
> at org.jgroups.protocols.FlowControl.up(FlowControl.java:381)
> at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:435)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75)
> at org.jgroups.protocols.TP.passMessageUp(TP.java:1567)
> at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1783)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> -----------------------------------------------------------------------
> Found suspect string in log4j at line 2818
> [error 2016/06/17 09:20:10.116 UTC <Geode UDP INT-1,asf905-22535> tid=0x42]
> Exception deserializing message payload: [dst: 67.195.81.149<v0>:1025, src:
> 67.195.81.149<v5>:1030 (2 headers), size=189 bytes,
> flags=OOB|NO_RELIABILITY|INTERNAL]
> java.io.IOException: Unknown header byte: 0
> at
> com.gemstone.gemfire.internal.InternalDataSerializer.basicReadObject(InternalDataSerializer.java:2979)
> at
> com.gemstone.gemfire.DataSerializer.readObject(DataSerializer.java:3271)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger.readJGMessage(JGroupsMessenger.java:849)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1043)
> at org.jgroups.JChannel.invokeCallback(JChannel.java:816)
> at org.jgroups.JChannel.up(JChannel.java:741)
> at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030)
> at org.jgroups.protocols.FRAG2.up(FRAG2.java:165)
> at org.jgroups.protocols.FlowControl.up(FlowControl.java:381)
> at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:435)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75)
> at
> com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75)
> at org.jgroups.protocols.TP.passMessageUp(TP.java:1567)
> at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1783)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> -----------------------------------------------------------------------
> Found suspect string in log4j at line 2858
> [fatal 2016/06/17 09:20:14.091 UTC <Test worker> tid=0x13] Rejected new
> system node 67.195.81.149(2109)<v5>:1030 because mcast was enabled which does
> not match the distributed system it is attempting to join. To fix this make
> sure the "mcast-port" gemfire property is set the same on all members of the
> same distributed system.
> com.gemstone.gemfire.IncompatibleSystemException: Rejected new system node
> 67.195.81.149(2109)<v5>:1030 because mcast was enabled which does not match
> the distributed system it is attempting to join. To fix this make sure the
> "mcast-port" gemfire property is set the same on all members of the same
> distributed system.
> at
> com.gemstone.gemfire.distributed.internal.DistributionManager.sendStartupMessage(DistributionManager.java:2678)
> at
> com.gemstone.gemfire.distributed.internal.DistributionManager.create(DistributionManager.java:557)
> at
> com.gemstone.gemfire.distributed.internal.InternalDistributedSystem.initialize(InternalDistributedSystem.java:616)
> at
> com.gemstone.gemfire.distributed.internal.InternalDistributedSystem.newInstance(InternalDistributedSystem.java:241)
> at
> com.gemstone.gemfire.distributed.DistributedSystem.connect(DistributedSystem.java:238)
> at
> com.gemstone.gemfire.distributed.LocatorDUnitTest.testMultipleLocatorsRestartingAtSameTime(LocatorDUnitTest.java:1483)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> at
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> at
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> at
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> at
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> at
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> at org.junit.rules.RunRules.evaluate(RunRules.java:20)
> at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> at
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> at
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> at
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> at
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecuter.runTestClass(JUnitTestClassExecuter.java:112)
> at
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecuter.execute(JUnitTestClassExecuter.java:56)
> at
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassProcessor.processTestClass(JUnitTestClassProcessor.java:66)
> at
> org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:35)
> at
> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
> at
> org.gradle.messaging.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:32)
> at
> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:93)
> at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
> at
> org.gradle.api.internal.tasks.testing.worker.TestWorker.processTestClass(TestWorker.java:109)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:35)
> at
> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
> at
> org.gradle.messaging.remote.internal.hub.MessageHub$Handler.run(MessageHub.java:360)
> at
> org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:54)
> at
> org.gradle.internal.concurrent.StoppableExecutorImpl$1.run(StoppableExecutorImpl.java:40)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)