[
https://issues.apache.org/jira/browse/HBASE-1061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
schubert zhang updated HBASE-1061:
----------------------------------
Attachment: hbase-stephen-regionserver-nd4-rack0-cloud.log
I enbaled the DEBUG log, and the log file is attached.
The state of HDFS is OK,
Node Last Contact Admin State Size (GB) Used (%) Used
(%) Remaining (GB) Blocks
nd1-rack0-cloud 1 In Service 822.8 1.13
755.27 806
nd2-rack0-cloud 2 In Service 822.8 1.1
755.4 3204
nd3-rack0-cloud 1 In Service 822.8 1.17
755.14 1918
nd4-rack0-cloud 2 In Service 822.8 1.18
755.13 2325
The loader code like following code:
HTable table = new HTable(ConfInstance.instance(), tableName);
for(int i = 0; ;++i) {
Date start = new Date(date.getTime() -
random.nextInt(30*24*60*60) * 1000L);
Integer end = random.nextInt(900000) + 100000;
String imsi = TestAdd.imsiBegin + end.toString();
BatchUpdate update = new BatchUpdate(imsi +
start.getTime());
update.put("gdr:start_time",
TypeDefine.toBytes(start.getTime()));
update.put("gdr:msisdn", TypeDefine.toBytes(imsi));
update.put("gdr:sourceip",
TypeDefine.toBytes(random.nextInt()));
update.put("gdr:destip",
TypeDefine.toBytes(random.nextInt()));
update.put("gdr:gdrtype",
TypeDefine.toBytes(random.nextInt(10) + 1));
update.put("gdr:req_num",
TypeDefine.toBytes(random.nextInt(10) + 1));
update.put("gdr:apn",
TypeDefine.toBytes(TestAdd.apns[random.nextInt(2)]));
update.put("gdr:gdrtime",
TypeDefine.toBytes(random.nextInt(10000) + 1));
update.put("gdr:result",
TypeDefine.toBytes(random.nextInt(2)));
update.put("gdr:gtpver",
TypeDefine.toBytes(random.nextInt(100) + 1));
update.put("gdr:remoteno",
TypeDefine.toBytes(random.nextInt(1000) + 1));
update.put("gdr:frontno",
TypeDefine.toBytes(random.nextInt(1000) + 1));
update.put("gdr:url",
TypeDefine.toBytes(TestAdd.urls[random.nextInt(TestAdd.urls.length)]));
try {
table.commit(update);
} catch (Exception e) {
e.printStackTrace();
Thread.sleep(60000);
}
> Region Server throw NotServingRegionException when batchUpdate during
> compaction/spliting and then shutdown
> ------------------------------------------------------------------------------------------------------------
>
> Key: HBASE-1061
> URL: https://issues.apache.org/jira/browse/HBASE-1061
> Project: Hadoop HBase
> Issue Type: Bug
> Components: regionserver
> Affects Versions: 0.18.1
> Environment: 1 master, 4 region server
> Hadoop 0.18.1
> When about 30GB 30,000,000 rows loaded
> Reporter: schubert zhang
> Attachments: hbase-stephen-regionserver-nd4-rack0-cloud.log
>
>
> 2008-12-13 21:30:58,184 INFO org.apache.hadoop.hbase.regionserver.LogRoller:
> Rolling hlog. Number of entries: 30015
> 2008-12-13 21:30:58,595 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> starting compaction on region GDR,138884579261228161625885,1229166027688
> 2008-12-13 21:31:00,162 INFO org.apache.hadoop.hbase.regionserver.HLog: New
> log writer created at
> /hbase/log_10.24.1.12_1229150189049_60020/hlog.dat.1229175060158
> 2008-12-13 21:31:00,162 INFO org.apache.hadoop.hbase.regionserver.HLog:
> removing old log file
> /hbase/log_10.24.1.12_1229150189049_60020/hlog.dat.1229173244235 whose
> highest sequence/edit id is 145008256
> 2008-12-13 21:31:00,165 INFO org.apache.hadoop.hbase.regionserver.HLog:
> removing old log file
> /hbase/log_10.24.1.12_1229150189049_60020/hlog.dat.1229173248573 whose
> highest sequence/edit id is 145038271
> 2008-12-13 21:31:05,571 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> compaction completed on region GDR,138884579261228161625885,1229166027688 in
> 6sec
> 2008-12-13 21:31:05,571 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> Starting split of region GDR,138884579261228161625885,1229166027688
> 2008-12-13 21:31:05,735 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> closed GDR,138884579261228161625885,1229166027688
> 2008-12-13 21:31:05,863 INFO org.apache.hadoop.ipc.Server: IPC Server handler
> 4 on 60020, call batchUpdate([...@4df611bd, row => 138884603141228853661885,
> {column => gdr:start_time, value => '...', column => gdr:imsi, value =>
> '...', column => gdr:msisdn, value => '...', column => gdr:sourceip, value =>
> '...', column => gdr:destip, value => '...', column => gdr:gdrtype, value =>
> '...', column => gdr:req_num, value => '...', column => gdr:apn, value =>
> '...', column => gdr:gdrtime, value => '...', column => gdr:result, value =>
> '...', column => gdr:gtpver, value => '...', column => gdr:remoteno, value =>
> '...', column => gdr:frontno, value => '...', column => gdr:offset, value =>
> '...', column => gdr:url, value => '...'}, -1) from 10.24.1.10:39000: error:
> org.apache.hadoop.hbase.NotServingRegionException: Region
> GDR,138884579261228161625885,1229166027688 closed
> org.apache.hadoop.hbase.NotServingRegionException: Region
> GDR,138884579261228161625885,1229166027688 closed
> at
> org.apache.hadoop.hbase.regionserver.HRegion.obtainRowLock(HRegion.java:1836)
> at
> org.apache.hadoop.hbase.regionserver.HRegion.getLock(HRegion.java:1901)
> at
> org.apache.hadoop.hbase.regionserver.HRegion.batchUpdate(HRegion.java:1432)
> at
> org.apache.hadoop.hbase.regionserver.HRegion.batchUpdate(HRegion.java:1406)
> at
> org.apache.hadoop.hbase.regionserver.HRegionServer.batchUpdate(HRegionServer.java:1415)
> at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:597)
> at org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:554)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:888)
> 2008-12-13 21:31:07,594 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> region GDR,138884579261228161625885,1229175065574/642186371 available
> 2008-12-13 21:31:07,594 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> closed GDR,138884579261228161625885,1229175065574
> 2008-12-13 21:31:07,885 INFO org.apache.hadoop.ipc.Server: IPC Server handler
> 9 on 60020, call batchUpdate([...@7792b8d, row => 138884603141228853661885,
> {column => gdr:start_time, value => '...', column => gdr:imsi, value =>
> '...', column => gdr:msisdn, value => '...', column => gdr:sourceip, value =>
> '...', column => gdr:destip, value => '...', column => gdr:gdrtype, value =>
> '...', column => gdr:req_num, value => '...', column => gdr:apn, value =>
> '...', column => gdr:gdrtime, value => '...', column => gdr:result, value =>
> '...', column => gdr:gtpver, value => '...', column => gdr:remoteno, value =>
> '...', column => gdr:frontno, value => '...', column => gdr:offset, value =>
> '...', column => gdr:url, value => '...'}, -1) from 10.24.1.10:39000: error:
> org.apache.hadoop.hbase.NotServingRegionException: Region
> GDR,138884579261228161625885,1229166027688 closed
> org.apache.hadoop.hbase.NotServingRegionException: Region
> GDR,138884579261228161625885,1229166027688 closed
> at
> org.apache.hadoop.hbase.regionserver.HRegion.obtainRowLock(HRegion.java:1836)
> at
> org.apache.hadoop.hbase.regionserver.HRegion.getLock(HRegion.java:1901)
> at
> org.apache.hadoop.hbase.regionserver.HRegion.batchUpdate(HRegion.java:1432)
> at
> org.apache.hadoop.hbase.regionserver.HRegion.batchUpdate(HRegion.java:1406)
> at
> org.apache.hadoop.hbase.regionserver.HRegionServer.batchUpdate(HRegionServer.java:1415)
> at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:597)
> at org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:554)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:888)
> 2008-12-13 21:31:08,346 INFO org.apache.hadoop.hbase.regionserver.HRegion:
> region GDR,138884596811227665528885,1229175065574/1240335718 available
> The many region throw such exception. and then the region server shutdown
> 2008-12-13 22:01:46,700 WARN org.apache.hadoop.dfs.DFSClient: DataStreamer
> Exception: java.io.IOException: Unable to create new block.
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2349)
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1800(DFSClient.java:1735)
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:1912)
> 2008-12-13 22:01:46,701 WARN org.apache.hadoop.dfs.DFSClient: Error Recovery
> for block blk_-1991134958879829135_36755 bad datanode[1]
> 2008-12-13 22:01:46,701 FATAL org.apache.hadoop.hbase.regionserver.Flusher:
> Replay of hlog required. Forcing server shutdown
> and at last:
> 008-12-13 22:04:53,414 INFO org.apache.hadoop.ipc.Client: Retrying connect to
> server: /10.24.1.18:60020. Already tried 6 time(s).
> 2008-12-13 22:04:54,417 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: /10.24.1.18:60020. Already tried 7 time(s).
> 2008-12-13 22:04:55,420 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: /10.24.1.18:60020. Already tried 8 time(s).
> 2008-12-13 22:04:56,423 INFO org.apache.hadoop.ipc.Client: Retrying connect
> to server: /10.24.1.18:60020. Already tried 9 time(s).
> 2008-12-13 22:07:04,556 ERROR
> org.apache.hadoop.hbase.regionserver.CompactSplitThread: Compaction failed
> for region GDR,138883452271227827193885,1229167157710
> java.lang.NullPointerException
> at
> org.apache.hadoop.hbase.client.ServerCallable.getRegionName(ServerCallable.java:71)
> at
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.getRegionServerWithRetries(HConnectionManager.java:863)
> at org.apache.hadoop.hbase.client.HTable.commit(HTable.java:964)
> at org.apache.hadoop.hbase.client.HTable.commit(HTable.java:950)
> at
> org.apache.hadoop.hbase.regionserver.CompactSplitThread.split(CompactSplitThread.java:167)
> at
> org.apache.hadoop.hbase.regionserver.CompactSplitThread.run(CompactSplitThread.java:85)
> 2008-12-13 22:07:04,558 INFO
> org.apache.hadoop.hbase.regionserver.CompactSplitThread:
> regionserver/0:0:0:0:0:0:0:0:60020.compactor exiting
> 2008-12-13 22:07:04,558 INFO
> org.apache.hadoop.hbase.regionserver.HRegionServer:
> regionserver/0:0:0:0:0:0:0:0:60020 exiting
> 2008-12-13 22:07:04,627 INFO
> org.apache.hadoop.hbase.regionserver.HRegionServer: Starting shutdown thread.
> 2008-12-13 22:07:04,627 INFO
> org.apache.hadoop.hbase.regionserver.HRegionServer: Shutdown thread complete
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.