How often do you see this error ? Is it possible to encapsulate the scenario using unit test ?
I am not certain HBASE-11234 On Jul 24, 2014, at 9:37 PM, "[email protected]" <[email protected]> wrote: > i use cdh hbase > 0.96.1.1,http://archive.cloudera.com/cdh5/cdh/5/hbase-0.96.1.1-cdh5.0.2.releasenotes.html > > no this patch > > so i add this path and rebuid hbase,restart regionServer will solve ths > problem?don't need to empty the data > > thanks > > > [email protected] > > From: Ted Yu > Date: 2014-07-25 13:20 > To: [email protected] > Subject: Re: hbase server exception > The error indicates there was bug in prefix tree module. > > Does the 0.96 release you use include fix for HBASE-11234 ? > > Cheers > > > On Thu, Jul 24, 2014 at 9:11 PM, [email protected] <[email protected]> wrote: > >> my hbase version is 0.96,and regionServer throw this exception,is this a >> bug? >> >> 2014-07-25 13:04:07,494 ERROR >> [regionserver60020-largeCompactions-1406165462536] >> regionserver.CompactSplitThread: Compaction failed Request = >> regionName=t2013,15100405147|1386470415000|472507379770398,1406208607841.763f79d0ed1d836b5748b70bf8a918cc., >> storeName=b, fileCount=10, fileSize=10.3 G (6.7 G, 3.1 G, 432.1 M, 70.1 M, >> 7.7 M, 4.8 M, 7.7 M, 132.1 K, 7.4 M, 7.7 M), priority=2099999901, >> time=2069097954607173 >> java.lang.ArrayIndexOutOfBoundsException >> at org.apache.hadoop.hbase.CellUtil.copyValueTo(CellUtil.java:105) >> at >> org.apache.hadoop.hbase.KeyValueUtil.appendToByteArray(KeyValueUtil.java:115) >> at >> org.apache.hadoop.hbase.KeyValueUtil.copyToNewByteArray(KeyValueUtil.java:90) >> at >> org.apache.hadoop.hbase.KeyValueUtil.copyToNewKeyValue(KeyValueUtil.java:74) >> at >> org.apache.hadoop.hbase.codec.prefixtree.PrefixTreeSeeker.getKeyValue(PrefixTreeSeeker.java:98) >> at >> org.apache.hadoop.hbase.io.hfile.HFileReaderV2$EncodedScannerV2.getKeyValue(HFileReaderV2.java:1070) >> at >> org.apache.hadoop.hbase.io.HalfStoreFileReader$1.getKeyValue(HalfStoreFileReader.java:143) >> at >> org.apache.hadoop.hbase.regionserver.StoreFileScanner.seek(StoreFileScanner.java:150) >> at >> org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:240) >> at >> org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:200) >> at >> org.apache.hadoop.hbase.regionserver.compactions.Compactor.createScanner(Compactor.java:259) >> at >> org.apache.hadoop.hbase.regionserver.compactions.DefaultCompactor.compact(DefaultCompactor.java:64) >> at >> org.apache.hadoop.hbase.regionserver.DefaultStoreEngine$DefaultCompactionContext.compact(DefaultStoreEngine.java:103) >> at org.apache.hadoop.hbase.regionserver.HStore.compact(HStore.java:980) >> at org.apache.hadoop.hbase.regionserver.HRegion.compact(HRegion.java:1415) >> at >> org.apache.hadoop.hbase.regionserver.CompactSplitThread$CompactionRunner.run(CompactSplitThread.java:475) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> at java.lang.Thread.run(Thread.java:745) >> >> [email protected] >>
