This is likely a new bug. 

Please consider opening a JIRA. 

> On Jan 16, 2017, at 2:22 AM, Pankaj kr <pankaj...@huawei.com> wrote:
> 
> Yeah, HBASE-13123 already available.
> 
> While going through HFilePrettyPrinter result with -e option, we observed 
> some keys doesn't contain complete column_family:qualifier information. 
> Normally in HFIle Keys are formed as,
>    K: r1/CF:Qualifier/1484302946516/Put/vlen=1/seqid=0
> 
> But some records,
>    K: r1/Unknown_Qualifier/1484302946516/Put/vlen=1/seqid=0
> Here there is no column family detail, normally it is like column family name 
> and the colon (:)    .
> 
> It's strange, how key is written in wrong format in HFile.
> 
> 
> Regards,
> Pankaj
> 
> -----Original Message-----
> From: Ted Yu [mailto:yuzhih...@gmail.com] 
> Sent: Saturday, January 14, 2017 9:41 AM
> To: dev@hbase.apache.org
> Cc: u...@hbase.apache.org
> Subject: Re: Region comapction failed
> 
> w.r.t. #2, I did a quick search for bloom related fixes.
> 
> I found HBASE-13123 but it was in 1.0.2
> 
> Planning to spend more time in the next few days.
> 
>> On Fri, Jan 13, 2017 at 5:29 PM, Pankaj kr <pankaj...@huawei.com> wrote:
>> 
>> Thanks Ted for replying.
>> 
>> Actually issue happened in production environment and there are many 
>> HFiles in that store (can't get the file). As we don't log the file 
>> name which is corrupted, Is there anyway to get the corrupted  file name?
>> 
>> Block encoding is "NONE", table schema has bloom filter as "ROW", 
>> compression type is "Snappy" and durability is SKIP_WAL.
>> 
>> 
>> Regards,
>> Pankaj
>> 
>> 
>> -----Original Message-----
>> From: Ted Yu [mailto:yuzhih...@gmail.com]
>> Sent: Friday, January 13, 2017 10:30 PM
>> To: dev@hbase.apache.org
>> Cc: u...@hbase.apache.org
>> Subject: Re: Region comapction failed
>> 
>> In the second case, the error happened when writing hfile. Can you 
>> track down the path of the new file so that further investigation can be 
>> done ?
>> 
>> Does the table use any encoding ?
>> 
>> Thanks
>> 
>>> On Jan 13, 2017, at 2:47 AM, Pankaj kr <pankaj...@huawei.com> wrote:
>>> 
>>> Hi,
>>> 
>>> We met a weird issue in our production environment.
>>> 
>>> Region compaction is always failing with  following errors,
>>> 
>>> 1.
>>> 2017-01-10 02:19:10,427 | ERROR | regionserver/RS-HOST/RS-IP:
>> PORT-longCompactions-1483858654825 | Compaction failed Request = 
>> regionName=XXXX., storeName=XYZ, fileCount=6, fileSize=100.7 M (3.2 M, 
>> 20.8 M, 15.1 M, 20.9 M, 21.0 M, 19.7 M), priority=-5, 
>> time=1747414906352088 | 
>> org.apache.hadoop.hbase.regionserver.CompactSplitThread$
>> CompactionRunner.doCompaction(CompactSplitThread.java:562)
>>> java.io.IOException: ScanWildcardColumnTracker.checkColumn ran into 
>>> a
>> column actually smaller than the previous column:  XXXXXXX
>>>               at org.apache.hadoop.hbase.regionserver.
>> ScanWildcardColumnTracker.checkVersions(ScanWildcardColumnTracker.
>> java:114)
>>>               at org.apache.hadoop.hbase.regionserver.ScanQueryMatcher.
>> match(ScanQueryMatcher.java:457)
>>>               at org.apache.hadoop.hbase.regionserver.StoreScanner.
>> next(StoreScanner.java:551)
>>>               at org.apache.hadoop.hbase.regionserver.compactions.
>> Compactor.performCompaction(Compactor.java:328)
>>>               at org.apache.hadoop.hbase.regionserver.compactions.
>> DefaultCompactor.compact(DefaultCompactor.java:104)
>>>               at org.apache.hadoop.hbase.regionserver.
>> DefaultStoreEngine$DefaultCompactionContext.compact(DefaultStoreEngine.
>> java:133)
>>>               at 
>>> org.apache.hadoop.hbase.regionserver.HStore.compact(
>> HStore.java:1243)
>>>               at 
>>> org.apache.hadoop.hbase.regionserver.HRegion.compact(
>> HRegion.java:1895)
>>>               at org.apache.hadoop.hbase.regionserver.
>> CompactSplitThread$CompactionRunner.doCompaction(
>> CompactSplitThread.java:546)
>>>               at org.apache.hadoop.hbase.regionserver.
>> CompactSplitThread$CompactionRunner.run(CompactSplitThread.java:583)
>>>               at java.util.concurrent.ThreadPoolExecutor.runWorker(
>> ThreadPoolExecutor.java:1142)
>>>               at java.util.concurrent.ThreadPoolExecuto
>>> 
>>> 2.
>>> 2017-01-10 02:33:53,009 | ERROR | regionserver/RS-HOST/RS-IP:
>> PORT-longCompactions-1483686810953 | Compaction failed Request = 
>> regionName=YYYYYY, storeName=ABC, fileCount=6, fileSize=125.3 M (20.9 
>> M,
>> 20.9 M, 20.9 M, 20.9 M, 20.9 M, 20.9 M), priority=-68,
>> time=1748294500157323 | org.apache.hadoop.hbase.regionserver.
>> CompactSplitThread$CompactionRunner.doCompaction(
>> CompactSplitThread.java:562)
>>> java.io.IOException: Non-increasing Bloom keys: 
>>> XXXXXXXXXXXXXXXXXXXXXX
>> after XXXXXXXXXXXX
>>>               at org.apache.hadoop.hbase.regionserver.StoreFile$Writer.
>> appendGeneralBloomfilter(StoreFile.java:911)
>>>               at org.apache.hadoop.hbase.regionserver.StoreFile$Writer.
>> append(StoreFile.java:947)
>>>               at org.apache.hadoop.hbase.regionserver.compactions.
>> Compactor.performCompaction(Compactor.java:337)
>>>               at org.apache.hadoop.hbase.regionserver.compactions.
>> DefaultCompactor.compact(DefaultCompactor.java:104)
>>>               at org.apache.hadoop.hbase.regionserver.
>> DefaultStoreEngine$DefaultCompactionContext.compact(DefaultStoreEngine.
>> java:133)
>>>               at 
>>> org.apache.hadoop.hbase.regionserver.HStore.compact(
>> HStore.java:1243)
>>>               at 
>>> org.apache.hadoop.hbase.regionserver.HRegion.compact(
>> HRegion.java:1895)
>>>               at org.apache.hadoop.hbase.regionserver.
>> CompactSplitThread$CompactionRunner.doCompaction(
>> CompactSplitThread.java:546)
>>>               at org.apache.hadoop.hbase.regionserver.
>> CompactSplitThread$CompactionRunner.run(CompactSplitThread.java:583)
>>>               at java.util.concurrent.ThreadPoolExecutor.runWorker(
>> ThreadPoolExecutor.java:1142)
>>>               at 
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(
>> ThreadPoolExecutor.java:617)
>>>               at java.lang.Thread.run(Thread.java:745)
>>> 
>>> HBase version : 1.0.2
>>> 
>>> We have verified all the HFiles in the store using 
>>> HFilePrettyPrinter
>> with "k" (checkrow), all report is normal. Full scan is also successful.
>>> We don't have the access to the actual data and may be customer wont
>> agree to  share that .
>>> 
>>> Have anyone faced this issue, any pointers will be much appreciated.
>>> 
>>> Thanks & Regards,
>>> Pankaj
>> 

Reply via email to