Look at your datanode logs around the same time. You probably either have this
http://wiki.apache.org/hadoop/Hbase/Troubleshooting#A5 or that http://wiki.apache.org/hadoop/Hbase/FAQ#A6 Also you see to be putting a fair number of regions on those region servers judging by the metrics, do consider setting HBASE_HEAP higher than 1GB in conf/hbase-env.sh J-D On Mon, Apr 5, 2010 at 8:38 PM, steven zhuang <steven.zhuang.1...@gmail.com> wrote: > greetings, > > while I was importing data into my HBase Cluster, I found one > regionserver is down, and by check the log, I found following exceptoin: > *EOFException*(during HBase flush memstore to HDFS file? not sure) > > seems that it's caused by DFSClient not working, I don't know the > exact reason, maybe it's caused by the heavy load on the machine where the > datanode is residing on, or the disk is full. but I am not sure which DFS > node should I check. > has anybody met the same problem? any pointer or hint is > appreciated. > > The log is as follows: > > > 2010-04-06 03:04:34,065 INFO org.apache.hadoop.hbase.regionserver.HRegion: > Blocking updates for 'IPC Server handler 20 on 60020' on region > hbt2table16,,1270522012397: memstore size 128.0m is >= than blocking 128.0m > size > 2010-04-06 03:04:34,712 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Completed compaction of 34; new storefile is > hdfs://rra-03:8887hbase/hbt2table16/2144402082/34/854678344516838047; store > size is 2.9m > 2010-04-06 03:04:34,715 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Compaction size of 35: 2.9m; Skipped 0 file(s), size: 0 > 2010-04-06 03:04:34,715 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Started compaction of 5 file(s) into > hbase/hbt2table16/compaction.dir/2144402082, seqid=2914432737 > 2010-04-06 03:04:35,055 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Added > hdfs://rra-03:8887hbase/hbt2table16/2144402082/184/1530971405029654438, > entries=1489, sequenceid=2914917785, memsize=203.8k, filesize=88.6k to > hbt2table16,,1270522012397 > 2010-04-06 03:04:35,442 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Completed compaction of 35; new storefile is > hdfs://rra-03:8887hbase/hbt2table16/2144402082/35/2952180521700205032; store > size is 2.9m > 2010-04-06 03:04:35,445 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Compaction size of 36: 2.9m; Skipped 0 file(s), size: 0 > 2010-04-06 03:04:35,445 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Started compaction of 4 file(s) into > hbase/hbt2table16/compaction.dir/2144402082, seqid=2914432737 > 2010-04-06 03:04:35,469 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Added > hdfs://rra-03:8887hbase/hbt2table16/2144402082/185/1984548574711437130, > entries=2105, sequenceid=2914917785, memsize=286.7k, filesize=123.9k to > hbt2table16,,1270522012397 > 2010-04-06 03:04:35,711 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Added > hdfs://rra-03:8887hbase/hbt2table16/2144402082/186/2470661482474884005, > entries=3031, sequenceid=2914917785, memsize=414.0k, filesize=179.1k to > hbt2table16,,1270522012397 > 2010-04-06 03:04:35,866 DEBUG > org.apache.hadoop.hbase.io.hfile.LruBlockCache: Block cache LRU eviction > started. Attempting to free 20853136 bytes > 2010-04-06 03:04:37,010 DEBUG > org.apache.hadoop.hbase.io.hfile.LruBlockCache: Block cache LRU eviction > completed. Freed 20866928 bytes. Priority Sizes: Single=17.422821MB > (18269152), Multi=150.70126MB (158021728),Memory=0.0MB (0) > 2010-04-06 03:04:37,542 INFO org.apache.hadoop.hdfs.DFSClient: Exception in > createBlockOutputStream java.io.EOFException > 2010-04-06 03:04:37,542 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning > block blk_-6935524980745310745_1391901 > 2010-04-06 03:04:37,607 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Completed compaction of 36; new storefile is > hdfs://rra-03:8887hbase/hbt2table16/2144402082/36/1570089400510240916; store > size is 2.9m > 2010-04-06 03:04:37,612 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Compaction size of 37: 2.9m; Skipped 0 file(s), size: 0 > 2010-04-06 03:04:37,612 DEBUG org.apache.hadoop.hbase.regionserver.Store: > Started compaction of 4 file(s) into > hbase/hbt2table16/compaction.dir/2144402082, seqid=2914432737 > 2010-04-06 03:04:37,964 INFO org.apache.hadoop.hdfs.DFSClient: Exception in > createBlockOutputStream java.io.*EOFException* > 2010-04-06 03:04:37,964 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning > block blk_2467598422201289982_1391902 > 2010-04-06 03:04:43,568 INFO org.apache.hadoop.hdfs.DFSClient: Exception in > createBlockOutputStream java.io.EOFException > 2010-04-06 03:04:43,568 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning > block blk_-2065206049437531800_1391902 > 2010-04-06 03:04:44,044 INFO org.apache.hadoop.hdfs.DFSClient: Exception in > createBlockOutputStream java.io.EOFException > 2010-04-06 03:04:44,044 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning > block blk_-3059563223628992257_1391902 > 2010-04-06 03:05:01,588 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer > Exception: java.io.IOException: Unable to create new block. > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2814) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2078) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2264) > > 2010-04-06 03:05:01,588 WARN org.apache.hadoop.hdfs.DFSClient: Error > Recovery for block blk_-3320281088550177280_1391903 bad datanode[0] nodes == > null > 2010-04-06 03:05:01,589 WARN org.apache.hadoop.hdfs.DFSClient: Could not get > block locations. Source file > "hbase/hbt2table16/2144402082/187/6358539637638901699" - Aborting... > 2010-04-06 03:05:01,589 FATAL > org.apache.hadoop.hbase.regionserver.MemStoreFlusher: Replay of hlog > required. Forcing server shutdown > org.apache.hadoop.hbase.DroppedSnapshotException: region: > hbt2table16,,1270522012397 > at > org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:977) > at > org.apache.hadoop.hbase.regionserver.HRegion.flushcache(HRegion.java:846) > at > org.apache.hadoop.hbase.regionserver.MemStoreFlusher.flushRegion(MemStoreFlusher.java:241) > at > org.apache.hadoop.hbase.regionserver.MemStoreFlusher.run(MemStoreFlusher.java:149) > Caused by: java.io.EOFException > at java.io.DataInputStream.readByte(DataInputStream.java:250) > at org.apache.hadoop.io.WritableUtils.readVLong(WritableUtils.java:298) > at org.apache.hadoop.io.WritableUtils.readVInt(WritableUtils.java:319) > at org.apache.hadoop.io.Text.readString(Text.java:400) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.createBlockOutputStream(DFSClient.java:2870) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2795) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2078) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2264) > 2010-04-06 03:05:01,603 INFO > org.apache.hadoop.hbase.regionserver.HRegionServer: Dump of metrics: > request=0.0, regions=335, stores=590, storefiles=1231, > storefileIndexSize=83, memstoreSize=128, compactionQueueSize=1, > usedHeap=710, maxHeap=993, blockCacheSize=162178088, > blockCacheFree=46200184, blockCacheCount=2483, blockCacheHitRatio=2, > fsReadLatency=0, fsWriteLatency=0, fsSyncLatency=0 > 2010-04-06 03:05:01,604 INFO > org.apache.hadoop.hbase.regionserver.MemStoreFlusher: > regionserver/10.76.112.214:60020.cacheFlusher exiting > 2010-04-06 03:05:01,673 INFO org.apache.hadoop.hbase.regionserver.HLog: Roll > hbase/.logs/rrb-08.off.tn.ask.com,60020,1268973923999/hlog.dat.1270523052543, > entries=483321, calcsize=88970157, filesize=61838598. New hlog hbase/.logs/ > rrb-08.off.tn.ask.com,60020,126897392 >