Hello,
We’ve got problem with hbase:meta that cause several servers to die.
Our version is HBase-1.4.8
prod006 - flush hbase:meta and move it:
Aug 13 15:46:23 prod006 hbase[63435]: 2019-08-13 15:46:23,019 INFO
[prod006,60020,1564942121270_ChoreService_1] regionserver.HRegionServer:
prod006,60020,1564942121270-MemstoreFlusherChore requesting flush of
hbase:meta,,1.1588230740 because info has an old edit so flush to free WALs
after random delay 226,666 ms
Aug 13 15:50:09 prod006 hbase[63435]: 2019-08-13 15:50:09,709 INFO
[MemStoreFlusher.1] regionserver.HRegion: Finished memstore flush of ~2.99
KB/3064, currentsize=0 B/0 for region hbase:meta,,1.1588230740 in 24ms,
sequenceid=323883, compaction requested=false
Aug 13 15:58:10 prod006 hbase[63435]: 2019-08-13 15:58:10,943 INFO
[StoreCloserThread-hbase:namespace,,1516784579153.d3d9ad37f8506c02a84458121965c0b0.-1]
regionserver.HStore: Closed info
Aug 13 15:58:10 prod006 hbase[63435]: 2019-08-13 15:58:10,951 INFO
[StoreCloserThread-hbase:meta,,1.1588230740-1] regionserver.HStore: Closed info
Aug 13 15:58:10 prod006 hbase[63435]: 2019-08-13 15:58:10,976 INFO
[RS_CLOSE_META-prod006:60020-0] regionserver.HRegion: Closed
hbase:meta,,1.1588230740
Aug 13 15:58:10 prod006 hbase[63435]: 2019-08-13 15:58:10,979 INFO
[RS_CLOSE_REGION-prod006:60020-0] regionserver.HRegion: Closed
hbase:namespace,,1516784579153.d3d9ad37f8506c02a84458121965c0b0.
Aug 13 15:58:10 prod006 hbase[63435]: 2019-08-13 15:58:10,979 INFO
[RS_CLOSE_REGION-prod006:60020-0] regionserver.HRegionServer: Adding moved
region record: d3d9ad37f8506c02a84458121965c0b0 to prod023,60020,1565701089060
as of 147
At this time at master we have this:
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,982 INFO
[AM.ZK.Worker-pool5-t1878] master.RegionStates: Transition {1588230740
state=PENDING_CLOSE, ts=1565701090850, server=prod006,60020,1564942121270} to
{158823074
0 state=CLOSED, ts=1565701090982, server=prod006,60020,1564942121270}
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,982 INFO
[AM.-pool3-t178] master.AssignmentManager: Setting node as OFFLINED in
ZooKeeper for region {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY
=> '', ENDKEY => ''}
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,982 INFO
[AM.-pool3-t178] master.RegionStates: Transition {1588230740 state=CLOSED,
ts=1565701090982, server=prod006,60020,1564942121270} to {1588230740
state=OFFLINE, ts=1565701090982, server=prod006,60020,1564942121270}
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,984 INFO
[AM.ZK.Worker-pool5-t1879] master.RegionStates: Transition
{d3d9ad37f8506c02a84458121965c0b0 state=PENDING_CLOSE, ts=1565701090871,
server=prod006,60020,1564942121270} to {d3d9ad37f8506c02a84458121965c0b0
state=CLOSED, ts=1565701090984, server=prod006,60020,1564942121270}
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,984 INFO
[AM.-pool3-t178] master.AssignmentManager: Assigning hbase:meta,,1.1588230740
to prod023,60020,1565701089060
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,984 INFO
[AM.-pool3-t178] master.RegionStates: Transition {1588230740 state=OFFLINE,
ts=1565701090982, server=prod006,60020,1564942121270} to {1588230740
state=PENDING_OPEN, ts=1565701090984, server=prod023,60020,1565701089060}
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,985 INFO
[AM.-pool3-t179] master.AssignmentManager: Setting node as OFFLINED in
ZooKeeper for region {ENCODED => d3d9ad37f8506c02a84458121965c0b0, NAME =>
'hbase:namespace,,1516784579153.d3d9ad37f8506c02a84458121965c0b0.', STARTKEY =>
'', ENDKEY => ''}
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,985 INFO
[AM.-pool3-t179] master.RegionStates: Transition
{d3d9ad37f8506c02a84458121965c0b0 state=CLOSED, ts=1565701090984,
server=prod006,60020,1564942121270} to {d3d9ad37f8506c02a84458121965c0b0
state=OFFLINE, ts=1565701090985, server=prod006,60020,1564942121270}
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,986 INFO
[AM.-pool3-t179] master.AssignmentManager: Assigning
hbase:namespace,,1516784579153.d3d9ad37f8506c02a84458121965c0b0. to
prod023,60020,1565701089060
Aug 13 15:58:10 prod001 hbase[14953]: 2019-08-13 15:58:10,986 INFO
[AM.-pool3-t179] master.RegionStates: Transition
{d3d9ad37f8506c02a84458121965c0b0 state=OFFLINE, ts=1565701090985,
server=prod006,60020,1564942121270} to {d3d9ad37f8506c02a84458121965c0b0
state=PENDING_OPEN, ts=1565701090986, server=prod023,60020,1565701089060}
Aug 13 15:58:11 prod001 hbase[14953]: 2019-08-13 15:58:11,039 WARN
[AM.-pool3-t179] master.AssignmentManager: Failed assignment of
hbase:namespace,,1516784579153.d3d9ad37f8506c02a84458121965c0b0. to
prod023,60020,1565701089060, waiting a little before trying on the same region
server try=1 of 10
Aug 13 15:58:11 prod001 hbase[14953]:
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException:
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server
prod023,60020,1565701089060 is not running yet
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.regionserver.RSRpcServices.checkOpen(RSRpcServices.java:1320)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.regionserver.RSRpcServices.openRegion(RSRpcServices.java:1707)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$2.callBlockingMethod(AdminProtos.java:22737)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2380)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:297)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:277)
Aug 13 15:58:11 prod001 hbase[14953]: at
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
Aug 13 15:58:11 prod001 hbase[14953]: at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
Aug 13 15:58:11 prod001 hbase[14953]: at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
Aug 13 15:58:11 prod001 hbase[14953]: at
java.lang.reflect.Constructor.newInstance(Constructor.java:423)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.RemoteWithExtrasException.instantiateException(RemoteWithExtrasException.java:95)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.RemoteWithExtrasException.unwrapRemoteException(RemoteWithExtrasException.java:85)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.protobuf.ProtobufUtil.makeIOExceptionOfException(ProtobufUtil.java:368)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:330)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.master.ServerManager.sendRegionOpen(ServerManager.java:787)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:2248)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1694)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.master.AssignCallable.call(AssignCallable.java:48)
Aug 13 15:58:11 prod001 hbase[14953]: at
java.util.concurrent.FutureTask.run(FutureTask.java:266)
Aug 13 15:58:11 prod001 hbase[14953]: at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
Aug 13 15:58:11 prod001 hbase[14953]: at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
Aug 13 15:58:11 prod001 hbase[14953]: at
java.lang.Thread.run(Thread.java:748)
Aug 13 15:58:11 prod001 hbase[14953]: Caused by:
org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.ipc.ServerNotRunningYetException):
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server
prod023,60020,1565701089060 is not running yet
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.regionserver.RSRpcServices.checkOpen(RSRpcServices.java:1320)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.regionserver.RSRpcServices.openRegion(RSRpcServices.java:1707)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$2.callBlockingMethod(AdminProtos.java:22737)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2380)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:297)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:277)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:386)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.AbstractRpcClient.access$100(AbstractRpcClient.java:94)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:409)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:405)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.Call.callComplete(Call.java:103)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.Call.setException(Call.java:118)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.BlockingRpcConnection.readResponse(BlockingRpcConnection.java:600)
Aug 13 15:58:11 prod001 hbase[14953]: at
org.apache.hadoop.hbase.ipc.BlockingRpcConnection.run(BlockingRpcConnection.java:334)
Aug 13 15:58:11 prod001 hbase[14953]: ... 1 more
Aug 13 15:58:11 prod001 hbase[14953]: 2019-08-13 15:58:11,217 INFO
[AM.ZK.Worker-pool5-t1882] master.RegionStates: Transition {1588230740
state=PENDING_OPEN, ts=1565701090984, server=prod023,60020,1565701089060} to
{1588230740 state=OPENING, ts=1565701091217, server=prod023,60020,1565701089060}
Aug 13 15:58:11 prod001 hbase[14953]: 2019-08-13 15:58:11,244 INFO
[AM.-pool3-t179] master.AssignmentManager: Assigning
hbase:namespace,,1516784579153.d3d9ad37f8506c02a84458121965c0b0. to
prod023,60020,1565701089060
Aug 13 15:58:11 prod001 hbase[14953]: 2019-08-13 15:58:11,391 INFO
[AM.ZK.Worker-pool5-t1883] master.RegionStates: Transition
{d3d9ad37f8506c02a84458121965c0b0 state=PENDING_OPEN, ts=1565701091244,
server=prod023,60020,1565701089060} to {d3d9ad37f8506c02a84458121965c0b0
state=OPENING, ts=1565701091391, server=prod023,60020,1565701089060}
Aug 13 15:58:11 prod001 hbase[14953]: 2019-08-13 15:58:11,666 INFO
[AM.ZK.Worker-pool5-t1884] master.RegionStates: Transition {1588230740
state=OPENING, ts=1565701091217, server=prod023,60020,1565701089060} to
{1588230740 state=OPEN, ts=1565701091666, server=prod023,60020,1565701089060}
Aug 13 15:58:11 prod001 hbase[14953]: 2019-08-13 15:58:11,666 INFO
[AM.ZK.Worker-pool5-t1884] coordination.ZkOpenRegionCoordination: Handling
OPENED of 1588230740 from prod001,60000,1565179968855; deleting unassigned node
Aug 13 15:58:11 prod001 hbase[14953]: 2019-08-13 15:58:11,668 INFO
[AM.ZK.Worker-pool5-t1886] master.RegionStates: Offlined 1588230740 from
prod006,60020,1564942121270
Aug 13 15:58:11 prod001 hbase[14953]: 2019-08-13 15:58:11,805 INFO
[AM.ZK.Worker-pool5-t1887] master.RegionStates: Transition
{d3d9ad37f8506c02a84458121965c0b0 state=OPENING, ts=1565701091391,
server=prod023,60020,1565701089060} to {d3d9ad37f8506c02a84458121965c0b0
state=OPEN, ts=1565701091805, server=prod023,60020,1565701089060}
Aug 13 15:58:11 prod001 hbase[14953]: 2019-08-13 15:58:11,807 INFO
[AM.ZK.Worker-pool5-t1889] master.RegionStates: Offlined
d3d9ad37f8506c02a84458121965c0b0 from prod006,60020,1564942121270
From this moment hbase:meta and hbase:namespace is moved to prod023 from prod006
But some other servers still try to get hbase:meta from prod006:
Aug 13 16:02:24 prod021 hbase[103657]: 2019-08-13 16:02:24,832 INFO
[regionserver/prod021/10.73.65.28:60020-splits-1564981231292]
client.RpcRetryingCaller: Call exception, tries=10, retries=350, started=38462
ms ago, cancelled=false, msg=org.apache.hadoop.hbase.NotServingRegionException:
Region hbase:meta,,1 is not online on prod006,60020,1564942121270
Aug 13 16:02:24 prod021 hbase[103657]: #011at
org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:3086)
Aug 13 16:02:24 prod021 hbase[103657]: #011at
org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1271)
Aug 13 16:02:24 prod021 hbase[103657]: #011at
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2197)
Aug 13 16:02:24 prod021 hbase[103657]: #011at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36617)
Aug 13 16:02:24 prod021 hbase[103657]: #011at
org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2369)
Aug 13 16:02:24 prod021 hbase[103657]: #011at
org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
Aug 13 16:02:24 prod021 hbase[103657]: #011at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:297)
Aug 13 16:02:24 prod021 hbase[103657]: #011at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:277)
Aug 13 16:02:24 prod021 hbase[103657]: row
‘IDX,\x1Abb6f7662-275e-459d-b0f3-7fd6aac9b6a4\x0000000046218537kV0WbJ_\x00\x13\x80\x00\x01k\x90\xFEzH\x00\x00\x00\x00,1562868892011.a215bcc7b396e5d08563f0a4183ba5cc.44'
on table 'hbase:meta' at region=hbase:meta,,1.1588230740,
hostname=prod023,60020,1565701089060, seqNum=0
….
Aug 13 16:21:41 prod021 hbase[103657]: 2019-08-13 16:21:41,136 INFO
[regionserver/prod021/10.73.65.28:60020-splits-1564981231292]
client.RpcRetryingCaller: Call exception, tries=68, retries=350,
started=1194772 ms ago, cancell
ed=false, msg=org.apache.hadoop.hbase.NotServingRegionException: Region
hbase:meta,,1 is not online on prod006,60020,1564942121270
Aug 13 16:21:41 prod021 hbase[103657]: 2019-08-13 16:21:41,146 INFO
[regionserver/prod021/10.73.65.28:60020-splits-1564981231292]
regionserver.SplitRequest: Running rollback/cleanup of failed split of
IDX,\x1Abb6f7662-275e-459d-b0f3-7fd6aac9b6a4\x0000000046218537kV0WbJ_\x00\x13\x80\x00\x01k\x90\xFEzH\x00\x00\x00\x00,1562868892011.a215bcc7b396e5d08563f0a4183ba5cc.;
callTimeout=1200000, callDuration=1214938:
org.apache.hadoop.hbase.NotServingRegionException: Region hbase:meta,,1 is not
online on prod006,60020,1564942121270
Aug 13 16:21:41 prod021 hbase[103657]: 2019-08-13 16:21:41,193 FATAL
[regionserver/prod021/10.73.65.28:60020-splits-1564981231292]
regionserver.HRegionServer: ABORTING region server prod021,60020,1564943354635:
Abort; we got an error after point-of-no-return
The same picture on other crashed servers
Any ideas why this happened and how to prevent this in future?