[
https://issues.apache.org/jira/browse/HBASE-24595?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17142683#comment-17142683
]
Yu Wang edited comment on HBASE-24595 at 6/23/20, 7:15 AM:
-----------------------------------------------------------
The phenomenon is similar with HBASE-22665 and regionserver log has the same
error log,but not found 'AbstractFSWAL.shutdown' in regionserver jstack.
the regionserver log has error log:
{code:java}
// Some comments here
public String getFoo()
{
return foo;
}
{code}
2020-06-23 14:34:11,943 ERROR [MemStoreFlusher.0] regionserver.MemStoreFlusher:
Cache flush failed for region hbase:meta,,1
org.apache.hadoop.hbase.exceptions.TimeoutIOException: Failed to get sync
result after 300000 ms for txid=22, WAL system stuck?
at
org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:145)
at
org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:718)
at
org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.sync(AsyncFSWAL.java:586)
at
org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2674)
at
org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2612)
at
org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2470)
at
org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2444)
at
org.apache.hadoop.hbase.regionserver.HRegion.flushcache(HRegion.java:2334)
at
org.apache.hadoop.hbase.regionserver.MemStoreFlusher.flushRegion(MemStoreFlusher.java:612)
at
org.apache.hadoop.hbase.regionserver.MemStoreFlusher.flushRegion(MemStoreFlusher.java:581)
at
org.apache.hadoop.hbase.regionserver.MemStoreFlusher.access$1000(MemStoreFlusher.java:68)
at
org.apache.hadoop.hbase.regionserver.MemStoreFlusher$FlushHandler.run(MemStoreFlusher.java:361)
at java.lang.Thread.run(Thread.java:748)
2020-06-23 14:34:35,011 WARN
[RpcServer.priority.FPBQ.Fifo.handler=17,queue=1,port=16020] ipc.RpcServer: Can
not complete this request in time, drop it: callId: 28 service: ClientService
methodName: Mutate size: 142 connection: 172.20.100.7:50198 deadline:
1592893834970 param: region=
hbase:namespace,,1592548148073.f5c7e71fb5e5cab3b27e52600996f7fd., row=aa1
connection: 172.20.100.7:50198
was (Author: [email protected]):
The phenomenon is similar with HBASE-22665 and regionserver log has the same
error log,but not found 'AbstractFSWAL.shutdown' in regionserver jstack.
the regionserver log has error log:
2020-06-23 14:34:11,943 ERROR [MemStoreFlusher.0] regionserver.MemStoreFlusher:
Cache flush failed for region hbase:meta,,1
org.apache.hadoop.hbase.exceptions.TimeoutIOException: Failed to get sync
result after 300000 ms for txid=22, WAL system stuck?
at
org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:145)
at
org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:718)
at
org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.sync(AsyncFSWAL.java:586)
at
org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2674)
at
org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2612)
at
org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2470)
at
org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2444)
at
org.apache.hadoop.hbase.regionserver.HRegion.flushcache(HRegion.java:2334)
at
org.apache.hadoop.hbase.regionserver.MemStoreFlusher.flushRegion(MemStoreFlusher.java:612)
at
org.apache.hadoop.hbase.regionserver.MemStoreFlusher.flushRegion(MemStoreFlusher.java:581)
at
org.apache.hadoop.hbase.regionserver.MemStoreFlusher.access$1000(MemStoreFlusher.java:68)
at
org.apache.hadoop.hbase.regionserver.MemStoreFlusher$FlushHandler.run(MemStoreFlusher.java:361)
at java.lang.Thread.run(Thread.java:748)
2020-06-23 14:34:35,011 WARN
[RpcServer.priority.FPBQ.Fifo.handler=17,queue=1,port=16020] ipc.RpcServer: Can
not complete this request in time, drop it: callId: 28 service: ClientService
methodName: Mutate size: 142 connection: 172.20.100.7:50198 deadline:
1592893834970 param: region=
hbase:namespace,,1592548148073.f5c7e71fb5e5cab3b27e52600996f7fd., row=aa1
connection: 172.20.100.7:50198
> hbase create namespace blocked when all datanodes has restarted
> ---------------------------------------------------------------
>
> Key: HBASE-24595
> URL: https://issues.apache.org/jira/browse/HBASE-24595
> Project: HBase
> Issue Type: Bug
> Affects Versions: 2.1.6
> Reporter: Yu Wang
> Priority: Critical
> Attachments: create_namespace_1.png, create_namespace_2.png,
> hmaster.log, hmaster.png, hmaster_4569.jstack, hregionserver.log,
> hregionserver_25649.jstack, procedure.png
>
>
> environment:
> jdk: 1.8.0_181
> hadoop: 3.1.1
> hbase: 2.1.6
> hbase shell create namespace blocked when all datanodes has restarted
> in kerberos environment,
> but create it successfully without kerberos
>
> hmaster日志中显示:
> 2020-06-19 23:47:48,241 WARN [PEWorker-15]
> procedure.CreateNamespaceProcedure: Retriable error trying to create
> namespace=abcd2 (in state=CREATE_NAMESPACE_INSERT_INTO_NS_TABLE)
> java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1220061:
> Call to hadoop-hbnn0005.com/172.20.101.36:16020 failed on local exception:
> org.apache.hadoop.hbase.ipc.CallTimeoutException: Call id=116,
> waitTime=10763, rpcTimeout=10759 row 'abcd2' on table 'hbase:namespace' at
> region=hbase:namespace,,1592548148073.f5c7e71fb5e5cab3b27e52600996f7fd.,
> hostname=hadoop-hbnn0005.com,16020,1592580274989, seqNum=162
> at
> org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithRetries(RpcRetryingCallerImpl.java:159)
> at org.apache.hadoop.hbase.client.HTable.put(HTable.java:542)
> at
> org.apache.hadoop.hbase.master.TableNamespaceManager.insertIntoNSTable(TableNamespaceManager.java:167)
> at
> org.apache.hadoop.hbase.master.procedure.CreateNamespaceProcedure.insertIntoNSTable(CreateNamespaceProcedure.java:240)
> at
> org.apache.hadoop.hbase.master.procedure.CreateNamespaceProcedure.executeFromState(CreateNamespaceProcedure.java:85)
> at
> org.apache.hadoop.hbase.master.procedure.CreateNamespaceProcedure.executeFromState(CreateNamespaceProcedure.java:39)
> at
> org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:189)
> at
> org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:965)
> at
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1723)
> at
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.executeProcedure(ProcedureExecutor.java:1462)
> at
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$1200(ProcedureExecutor.java:78)
> at
> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2039)
> Caused by: org.apache.hadoop.hbase.ipc.CallTimeoutException: Call to
> hadoop-hbnn0005.com/172.20.101.36:16020 failed on local exception:
> org.apache.hadoop.hbase.ipc.CallTimeoutException: Call id=116,
> waitTime=10763, rpcTimeout=10759
> at org.apache.hadoop.hbase.ipc.IPCUtil.wrapException(IPCUtil.java:205)
> at
> org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:390)
> at
> org.apache.hadoop.hbase.ipc.AbstractRpcClient.access$100(AbstractRpcClient.java:95)
> at
> org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:410)
> at
> org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:406)
> at org.apache.hadoop.hbase.ipc.Call.setTimeout(Call.java:96)
> at
> org.apache.hadoop.hbase.ipc.RpcConnection$1.run(RpcConnection.java:199)
> at
> org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$HashedWheelTimeout.expire(HashedWheelTimer.java:682)
> at
> org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$HashedWheelBucket.expireTimeouts(HashedWheelTimer.java:757)
> at
> org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:485)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: org.apache.hadoop.hbase.ipc.CallTimeoutException: Call id=116,
> waitTime=10763, rpcTimeout=10759
> at
> org.apache.hadoop.hbase.ipc.RpcConnection$1.run(RpcConnection.java:200)
> ... 4 more
> 2020-06-19 23:47:49,218 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 1.262sec
> 2020-06-19 23:47:54,220 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 6.263sec
> 2020-06-19 23:47:59,220 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 11.264sec
> 2020-06-19 23:48:04,220 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 16.264sec
> 2020-06-19 23:48:09,221 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 21.265sec
> 2020-06-19 23:48:14,221 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 26.265sec
> 2020-06-19 23:48:19,221 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 31.265sec
> 2020-06-19 23:48:24,222 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 36.266sec
> 2020-06-19 23:48:29,222 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 41.266sec
> 2020-06-19 23:48:34,223 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 46.267sec
> 2020-06-19 23:48:39,223 WARN [ProcExecTimeout] procedure2.ProcedureExecutor:
> Worker stuck PEWorker-15(pid=171), run time 20mins, 51.267sec
--
This message was sent by Atlassian Jira
(v8.3.4#803005)