Bruce Schuchardt created GEODE-6823:
---------------------------------------
Summary: Hang in ElderInitProcessor.init()
Key: GEODE-6823
URL: https://issues.apache.org/jira/browse/GEODE-6823
Project: Geode
Issue Type: Bug
Components: distributed lock service
Reporter: Bruce Schuchardt
A locator and a server were spinning up at the same time and the locator became
stuck trying to initialize a distributed lock service. Extra logging showed
that the server received an ElderInitMessage that it decided to ignore because
it thought it was shutting down.
{noformat}
gemfire2_2430/system.log: [info 2019/05/29 11:00:34.230 PDT
<vm_2_thr_40_client2_host1_2430> tid=0x24] Initial (distribution manager) view,
View[rs-GEM-2316-0906a2i3large-hydra-client-11(2354:locator)<ec><v0>:41000|2]
members:
[rs-GEM-2316-0906a2i3large-hydra-client-11(2354:locator)<ec><v0>:41000,
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire1_host1_2416:2416)<ec><v1>:41001{lead},
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire1_host1_2420:2420)<ec><v2>:41002,
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire2_host1_2430:2430)<ec><v2>:41003]
gemfire2_2430/system.log: [debug 2019/05/29 11:00:34.463 PDT <unicast
receiver,rs-GEM-2316-0906a2i3large-hydra-client-11-10705> tid=0x46] Received
message 'ElderInitMessage (processorId='1)' from
<rs-GEM-2316-0906a2i3large-hydra-client-11(2354:locator)<ec><v0>:41000>
gemfire2_2430/system.log: [debug 2019/05/29 11:00:34.574 PDT <Pooled Message
Processor 2> tid=0x4d] Waiting for Elder to change. Expecting Elder to be
rs-GEM-2316-0906a2i3large-hydra-client-11(2354:locator)<ec><v0>:41000, is
rs-GEM-2316-0906a2i3large-hydra-client-11(2354:locator)<ec><v0>:41000.
gemfire2_2430/system.log: [info 2019/05/29 11:00:34.575 PDT <Pooled Message
Processor 2> tid=0x4d] ElderInitMessage (processorId='1): disregarding request
from departed member.
gemfire2_2430/system.log: [info 2019/05/29 11:00:35.238 PDT <unicast
receiver,rs-GEM-2316-0906a2i3large-hydra-client-11-10705> tid=0x46] received
new view:
View[rs-GEM-2316-0906a2i3large-hydra-client-11(2354:locator)<ec><v0>:41000|3]
members:
[rs-GEM-2316-0906a2i3large-hydra-client-11(2354:locator)<ec><v0>:41000,
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire1_host1_2416:2416)<ec><v1>:41001{lead},
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire1_host1_2420:2420)<ec><v2>:41002,
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire2_host1_2430:2430)<ec><v2>:41003,
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire2_host1_2437:2437)<ec><v3>:41004]
locator_ds_2354/system.log: [warn 2019/05/29 11:00:50.430 PDT <Pooled Message
Processor 2> tid=0x38] 15 seconds have elapsed while waiting for replies:
<ElderInitProcessor 1 waiting for 1 replies from
[rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire2_host1_2430:2430)<ec><v2>:41003]>
on rs-GEM-2316-0906a2i3large-hydra-client-11(2354:locator)<ec><v0>:41000 whose
current membership list is:
[[rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire2_host1_2437:2437)<ec><v3>:41004,
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire2_host1_2430:2430)<ec><v2>:41003,
rs-GEM-2316-0906a2i3large-hydra-client-11(2354:locator)<ec><v0>:41000,
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire1_host1_2420:2420)<ec><v2>:41002,
rs-GEM-2316-0906a2i3large-hydra-client-11(gemfire1_host1_2416:2416)<ec><v1>:41001]]
[Stack #1 from bgexec15197_2354.log line 2]
"Pooled Message Processor 2" #56 daemon prio=5 os_prio=0 tid=0x000000000194e800
nid=0xae3 waiting on condition [0x00007f5c94dce000]
java.lang.Thread.State: TIMED_WAITING (parking)
at sun.misc.Unsafe.park(Native Method)
- parking to wait for <0x0000000775ff4f08> (a
java.util.concurrent.CountDownLatch$Sync)
at
java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedNanos(AbstractQueuedSynchronizer.java:1037)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1328)
at java.util.concurrent.CountDownLatch.await(CountDownLatch.java:277)
at
org.apache.geode.internal.util.concurrent.StoppableCountDownLatch.await(StoppableCountDownLatch.java:71)
at
org.apache.geode.distributed.internal.ReplyProcessor21.basicWait(ReplyProcessor21.java:716)
at
org.apache.geode.distributed.internal.ReplyProcessor21.waitForRepliesUninterruptibly(ReplyProcessor21.java:787)
at
org.apache.geode.distributed.internal.ReplyProcessor21.waitForRepliesUninterruptibly(ReplyProcessor21.java:764)
at
org.apache.geode.distributed.internal.ReplyProcessor21.waitForRepliesUninterruptibly(ReplyProcessor21.java:850)
at
org.apache.geode.distributed.internal.locks.ElderInitProcessor.init(ElderInitProcessor.java:69)
at
org.apache.geode.distributed.internal.locks.ElderState.<init>(ElderState.java:53)
at
org.apache.geode.distributed.internal.ClusterElderManager.lambda$new$0(ClusterElderManager.java:41)
at
org.apache.geode.distributed.internal.ClusterElderManager$$Lambda$64/1182435120.get(Unknown
Source)
at
org.apache.geode.distributed.internal.ClusterElderManager.initializeElderState(ClusterElderManager.java:107)
at
org.apache.geode.distributed.internal.ClusterElderManager.getElderState(ClusterElderManager.java:98)
at
org.apache.geode.distributed.internal.ClusterDistributionManager.getElderState(ClusterDistributionManager.java:2885)
at
org.apache.geode.distributed.internal.locks.GrantorRequestProcessor$GrantorRequestMessage.basicProcess(GrantorRequestProcessor.java:494)
at
org.apache.geode.distributed.internal.locks.GrantorRequestProcessor$GrantorRequestMessage.process(GrantorRequestProcessor.java:489)
at
org.apache.geode.distributed.internal.DistributionMessage.scheduleAction(DistributionMessage.java:369)
at
org.apache.geode.distributed.internal.DistributionMessage$1.run(DistributionMessage.java:435)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at
org.apache.geode.distributed.internal.ClusterDistributionManager.runUntilShutdown(ClusterDistributionManager.java:959)
at
org.apache.geode.distributed.internal.ClusterDistributionManager.doProcessingThread(ClusterDistributionManager.java:825)
at
org.apache.geode.distributed.internal.ClusterDistributionManager$$Lambda$77/482261794.invoke(Unknown
Source)
at
org.apache.geode.internal.logging.LoggingThreadFactory.lambda$newThread$0(LoggingThreadFactory.java:121)
at
org.apache.geode.internal.logging.LoggingThreadFactory$$Lambda$78/2015475584.run(Unknown
Source)
at java.lang.Thread.run(Thread.java:748)
so 2430 did not respond to the ElderInitMessage. That would happen if 2430 was
shutting down, if 2354 was no longer a member or if 2430 became the elder.
{noformat}
I added more logging to ClusterElderManager.waitForElder(), which is returning
false and causing this "disregarding" message to be logged. What I see is that
the server doing this thinks it's shutting down but it most definitely is not:
{noformat}
[debug 2019/05/29 15:58:48.488 PDT <unicast
receiver,rs-GEM-2316-1539a0i3large-hydra-client-4-39597> tid=0x46] Received
message 'ElderInitMessage (processorId=1)' from
<rs-GEM-2316-1539a0i3large-hydra-client-4(15224:locator)<ec><v0>:41000>
[debug 2019/05/29 15:58:48.510 PDT <Pooled Message Processor 2> tid=0x4d] Wait
for elder terminated because we're shutting down
[info 2019/05/29 15:58:48.511 PDT <Pooled Message Processor 2> tid=0x4d]
ElderInitMessage (processorId=1): disregarding request from departed member.
[debug 2019/05/29 15:58:48.599 PDT <unicast
receiver,rs-GEM-2316-1539a0i3large-hydra-client-4-39597> tid=0x46] processing
InstallViewMessage(type=PREPARE; Current ViewID=3; Previous View ID=0;
View[rs-GEM-2316-1539a0i3large-hydra-client-4(15224:locator)<ec><v0>:41000|3]
members:
[rs-GEM-2316-1539a0i3large-hydra-client-4(15224:locator)<ec><v0>:41000,
rs-GEM-2316-1539a0i3large-hydra-client-4(gemfire1_host1_15283:15283)<ec><v1>:41001{lead},
rs-GEM-2316-1539a0i3large-hydra-client-4(gemfire1_host1_15290:15290)<ec><v2>:41002,
rs-GEM-2316-1539a0i3large-hydra-client-4(gemfire2_host1_15295:15295)<ec><v3>:41004,
rs-GEM-2316-1539a0i3large-hydra-client-4(gemfire2_host1_15304:15304)<ec><v3>:41003];
cred=null){noformat}
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)