Mike Drob created HBASE-18628:
---------------------------------
Summary: ZKPermissionWatcher blocks all ZK notifications
Key: HBASE-18628
URL: https://issues.apache.org/jira/browse/HBASE-18628
Project: HBase
Issue Type: Bug
Components: regionserver
Reporter: Mike Drob
Priority: Critical
Buckle up folks, we're going for a ride here. I've seeing this on a branch-2
based build, but I think the problem will affect branch-1 as well. I'm not able
to easily reproduce the issue, but it will usually come up within an hour on a
given cluster that I have, at which point the problem persists until an RS
restart. I've been seeing the problem and paying attention for maybe two
months, but I suspect it's been happening much longer than that.
h3. Problem
When running in a secure cluster, sometimes the ZK EventThread will get stuck
on a permissions update and not be able to process new notifications. This
happens to also block flush and snapshot, which is how we found it.
h3. Analysis
The main smoking gun is seeing this in repeated jstacks:
{noformat}
"main-EventThread" #43 daemon prio=5 os_prio=0 tid=0x00007f0b92644000
nid=0x6e69 waiting on condition [0x00007f0b6730f000]
java.lang.Thread.State: TIMED_WAITING (sleeping)
at java.lang.Thread.sleep(Native Method)
at
org.apache.hadoop.hbase.security.access.ZKPermissionWatcher.nodeChildrenChanged(ZKPermissionWatcher.java:191)
at
org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher.process(ZooKeeperWatcher.java:503)
at
org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:522)
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:498)
{noformat}
That sleep is a 20ms sleep in an {{AtomicReference.compareAndSet}} loop - but
it never gets past the condition.
{code}
while (!nodes.compareAndSet(null, nodeList)) {
try {
Thread.sleep(20);
} catch (InterruptedException e) {
LOG.warn("Interrupted while setting node list", e);
Thread.currentThread().interrupt();
}
}
{code}
The warning never shows up in the logs, it just keeps looping and looping. The
last relevant line from the watcher in logs is:
{noformat}
2017-08-17 21:25:12,379 DEBUG
org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher:
regionserver:22101-0x15df38884c80024, quorum=zk1:2181,zk2:2181,zk3:2181,
baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged,
state=SyncConnected, path=/hbase/acl
{noformat}
Which makes sense, because the code snippet is from permission watcher's
{{nodeChildrenChanged}} handler.
The separate thread introduced in HBASE-14370 is present, but not doing
anything. And this event hasn't gotten to the part where it splits off into a
thread:
{noformat}
"zk-permission-watcher4-thread-1" #160 daemon prio=5 os_prio=0
tid=0x0000000001750800 nid=0x6fd9 waiting on condition [0x00007f0b5dce5000]
java.lang.Thread.State: WAITING (parking)
at sun.misc.Unsafe.park(Native Method)
- parking to wait for <0x00000007436ecea0> (a
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
{noformat}
h3. Solutions
There's a few approaches we can take to fix this, I think they are all
complimentary. It might be useful to file subtasks or new issues for some of
the solutions if they are longer term.
# Move flush and snapshot to ProcedureV2. This makes my proximate problem go
away, but it's only relevant to branch-2 and master, and doesn't fix anything
on branch-1. Also, Permissions updates would still get stuck, preventing future
permissions updates. I think this is important long term for the robustness of
the system, but not a viable short term fix.
# Add an Executor to ZookeeperWatcher and launch threads from there. Maybe we'd
want to pull the Executor out of ZKPW, but that's not strictly necessary and
can be optimized later -- if we're already threading, then adding another layer
isn't a huge cost.
# Figure out the race condition or logic problem that causes {{nodes}} to be
non-null above. I've tried looking at this and visual inspection isn't getting
me anywhere.
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)