I don't know if it is of any particular interest, but after a truncate I can use the table again.
The problem, of course, is that all data is gone. Lucas On Thu, Jul 2, 2009 at 11:40 AM, Lucas Nazário dos Santos < [email protected]> wrote: > Hi, > > It's the second time it happens. I have a Hadoop job that reads and inserts > data into HBase. It works perfectly for a couple of hours and then HBase > hangs. > > I'm using HBase 0.19.3 and Hadoop 0.19.1. > > Interesting is that the list command shows the table, but a count returns > an exception. > > Bellow is the error log. > > Does anybody know what is happening? > > Lucas > > > > 2009-07-02 05:38:02,337 INFO > org.apache.hadoop.hbase.regionserver.HRegionServer: MSG_REGION_FLUSH: > document,,1246496132379: safeMode=false > 2009-07-02 05:38:02,337 INFO > org.apache.hadoop.hbase.regionserver.HRegionServer: Worker: > MSG_REGION_FLUSH: document,,1246496132379: safeMode=false > 2009-07-02 05:40:11,518 WARN org.apache.hadoop.hdfs.DFSClient: Exception > while reading from blk_4097294633794140351_1008 of > /hbase/-ROOT-/70236052/info/mapfiles/6567566389528605238/index from > 192.168.1.3:50010: java.io.IOException: Premeture EOF from inputStream > at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:102) > at > org.apache.hadoop.hdfs.DFSClient$BlockReader.readChunk(DFSClient.java:1207) > at > org.apache.hadoop.fs.FSInputChecker.readChecksumChunk(FSInputChecker.java:238) > at org.apache.hadoop.fs.FSInputChecker.fill(FSInputChecker.java:177) > at org.apache.hadoop.fs.FSInputChecker.read1(FSInputChecker.java:194) > at org.apache.hadoop.fs.FSInputChecker.read(FSInputChecker.java:159) > at > org.apache.hadoop.hdfs.DFSClient$BlockReader.read(DFSClient.java:1060) > at > org.apache.hadoop.hdfs.DFSClient$DFSInputStream.readBuffer(DFSClient.java:1615) > at > org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1665) > at java.io.DataInputStream.readFully(DataInputStream.java:178) > at java.io.DataInputStream.readFully(DataInputStream.java:152) > at > org.apache.hadoop.hbase.io.SequenceFile$Reader.init(SequenceFile.java:1464) > at > org.apache.hadoop.hbase.io.SequenceFile$Reader.<init>(SequenceFile.java:1442) > at > org.apache.hadoop.hbase.io.SequenceFile$Reader.<init>(SequenceFile.java:1431) > at > org.apache.hadoop.hbase.io.SequenceFile$Reader.<init>(SequenceFile.java:1426) > at org.apache.hadoop.hbase.io.MapFile$Reader.open(MapFile.java:318) > at > org.apache.hadoop.hbase.io.HBaseMapFile$HBaseReader.<init>(HBaseMapFile.java:78) > at > org.apache.hadoop.hbase.io.BloomFilterMapFile$Reader.<init>(BloomFilterMapFile.java:68) > at > org.apache.hadoop.hbase.regionserver.HStoreFile.getReader(HStoreFile.java:443) > at > org.apache.hadoop.hbase.regionserver.StoreFileScanner.openReaders(StoreFileScanner.java:127) > at > org.apache.hadoop.hbase.regionserver.StoreFileScanner.<init>(StoreFileScanner.java:65) > at > org.apache.hadoop.hbase.regionserver.HStoreScanner.<init>(HStoreScanner.java:92) > at > org.apache.hadoop.hbase.regionserver.HStore.getScanner(HStore.java:2134) > at > org.apache.hadoop.hbase.regionserver.HRegion$HScanner.<init>(HRegion.java:2000) > at > org.apache.hadoop.hbase.regionserver.HRegion.getScanner(HRegion.java:1187) > at > org.apache.hadoop.hbase.regionserver.HRegionServer.openScanner(HRegionServer.java:1714) > at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:632) > at > org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:912) > > 2009-07-02 05:41:11,517 WARN org.apache.hadoop.hdfs.DFSClient: Exception > while reading from blk_4097294633794140351_1008 of > /hbase/-ROOT-/70236052/info/mapfiles/6567566389528605238/index from > 192.168.1.3:50010: java.io.IOException: Premeture EOF from inputStream > at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:102) > at > org.apache.hadoop.hdfs.DFSClient$BlockReader.readChunk(DFSClient.java:1207) > at > org.apache.hadoop.fs.FSInputChecker.readChecksumChunk(FSInputChecker.java:238) > at org.apache.hadoop.fs.FSInputChecker.fill(FSInputChecker.java:177) > at org.apache.hadoop.fs.FSInputChecker.read1(FSInputChecker.java:194) > at org.apache.hadoop.fs.FSInputChecker.read(FSInputChecker.java:159) > at > org.apache.hadoop.hdfs.DFSClient$BlockReader.read(DFSClient.java:1060) > at > org.apache.hadoop.hdfs.DFSClient$DFSInputStream.readBuffer(DFSClient.java:1615) > at > org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1665) > at java.io.DataInputStream.readFully(DataInputStream.java:178) > at java.io.DataInputStream.readFully(DataInputStream.java:152) > at > org.apache.hadoop.hbase.io.SequenceFile$Reader.init(SequenceFile.java:1464) > at > org.apache.hadoop.hbase.io.SequenceFile$Reader.<init>(SequenceFile.java:1442) > at > org.apache.hadoop.hbase.io.SequenceFile$Reader.<init>(SequenceFile.java:1431) > at > org.apache.hadoop.hbase.io.SequenceFile$Reader.<init>(SequenceFile.java:1426) > at org.apache.hadoop.hbase.io.MapFile$Reader.open(MapFile.java:318) > at > org.apache.hadoop.hbase.io.HBaseMapFile$HBaseReader.<init>(HBaseMapFile.java:78) > at > org.apache.hadoop.hbase.io.BloomFilterMapFile$Reader.<init>(BloomFilterMapFile.java:68) > at > org.apache.hadoop.hbase.regionserver.HStoreFile.getReader(HStoreFile.java:443) > at > org.apache.hadoop.hbase.regionserver.StoreFileScanner.openReaders(StoreFileScanner.java:127) > at > org.apache.hadoop.hbase.regionserver.StoreFileScanner.<init>(StoreFileScanner.java:65) > at > org.apache.hadoop.hbase.regionserver.HStoreScanner.<init>(HStoreScanner.java:92) > at > org.apache.hadoop.hbase.regionserver.HStore.getScanner(HStore.java:2134) > at > org.apache.hadoop.hbase.regionserver.HRegion$HScanner.<init>(HRegion.java:2000) > at > org.apache.hadoop.hbase.regionserver.HRegion.getScanner(HRegion.java:1187) > at > org.apache.hadoop.hbase.regionserver.HRegionServer.openScanner(HRegionServer.java:1714) > at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:632) > at > org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:912) > >
