[
https://issues.apache.org/jira/browse/HBASE-13962?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
reaz hedayati updated HBASE-13962:
----------------------------------
Description:
hi every body
my table has some cell that load with bulk load scenario and some cells for
increment.
we use 2 job to load data into table, first job use increment in reduce site
and second job use bulk load.
first we run increment job, next run bulk job and run completebulkload job,
after that we got this exception:
2015-06-24 17:40:01,557 INFO
[regionserver60020-smallCompactions-1434448531302] regionserver.HRegion:
Starting compaction on c2 in region table1,\x04C#P1"\x07\x94
,1435065082383.0fe38a6c782600e4d46f1f148144b489.
2015-06-24 17:40:01,558 INFO
[regionserver60020-smallCompactions-1434448531302] regionserver.HStore:
Starting compaction of 3 file(s) in c2 of table1,\x04C#P1"\x07\x94
,1435065082383.0fe38a6c782600e4d46f1f148144b489. into
tmpdir=hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/.tmp,
totalSize=43.1m
2015-06-24 17:40:01,558 DEBUG
[regionserver60020-smallCompactions-1434448531302] regionserver.StoreFileInfo:
reference
'hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/6b1249a3b474474db5cf6c664f2d98dc.d21f8ee8b3c915fd9e1c143a0f1892e5'
to region=d21f8ee8b3c915fd9e1c143a0f1892e5
hfile=6b1249a3b474474db5cf6c664f2d98dc
2015-06-24 17:40:01,558 DEBUG
[regionserver60020-smallCompactions-1434448531302] compactions.Compactor:
Compacting
hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/6b1249a3b474474db5cf6c664f2d98dc.d21f8ee8b3c915fd9e1c143a0f1892e5-hdfs://m2/hbase2/data/default/table1/d21f8ee8b3c915fd9e1c143a0f1892e5/c2/6b1249a3b474474db5cf6c664f2d98dc-top,
keycount=575485, bloomtype=ROW, size=20.8m, encoding=NONE, seqNum=9,
earliestPutTs=1434875448405
2015-06-24 17:40:01,558 DEBUG
[regionserver60020-smallCompactions-1434448531302] compactions.Compactor:
Compacting
hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/41e13b20ee79435ebc260d11d3bf9920_SeqId_11_,
keycount=562988, bloomtype=ROW, size=10.1m, encoding=NONE, seqNum=11,
earliestPutTs=1435076732205
2015-06-24 17:40:01,558 DEBUG
[regionserver60020-smallCompactions-1434448531302] compactions.Compactor:
Compacting
hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/565c45ff05b14a419978834c86defa1a_SeqId_12_,
keycount=554577, bloomtype=ROW, size=12.2m, encoding=NONE, seqNum=12,
earliestPutTs=1435136926850
2015-06-24 17:40:01,560 ERROR
[regionserver60020-smallCompactions-1434448531302]
regionserver.CompactSplitThread: Compaction failed Request =
regionName=table1,\x04C#P1"\x07\x94
,1435065082383.0fe38a6c782600e4d46f1f148144b489., storeName=c2, fileCount=3,
fileSize=43.1m (20.8m, 10.1m, 12.2m), priority=1, time=6077271921381072
java.io.IOException: Could not seek
StoreFileScanner[org.apache.hadoop.hbase.io.HalfStoreFileReader$1@1d1eb574,
cur=null] to key /c2:/LATEST_TIMESTAMP/DeleteFamily/vlen=0/mvcc=0
at
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seek(StoreFileScanner.java:164)
at
org.apache.hadoop.hbase.regionserver.StoreScanner.seekScanners(StoreScanner.java:329)
at
org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:252)
at
org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:214)
at
org.apache.hadoop.hbase.regionserver.compactions.Compactor.createScanner(Compactor.java:299)
at
org.apache.hadoop.hbase.regionserver.compactions.DefaultCompactor.compact(DefaultCompactor.java:87)
at
org.apache.hadoop.hbase.regionserver.DefaultStoreEngine$DefaultCompactionContext.compact(DefaultStoreEngine.java:112)
at org.apache.hadoop.hbase.regionserver.HStore.compact(HStore.java:1113)
at
org.apache.hadoop.hbase.regionserver.HRegion.compact(HRegion.java:1519)
at
org.apache.hadoop.hbase.regionserver.CompactSplitThread$CompactionRunner.run(CompactSplitThread.java:498)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: Failed to read compressed block at 10930320,
onDiskSizeWithoutHeader=22342, preReadHeaderSize=33, header.length=33, header
bytes:
\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00
at
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderV2.readBlockDataInternal(HFileBlock.java:1549)
at
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderV2.readBlockData(HFileBlock.java:1413)
at
org.apache.hadoop.hbase.io.hfile.HFileReaderV2.readBlock(HFileReaderV2.java:394)
at
org.apache.hadoop.hbase.io.hfile.HFileBlockIndex$BlockIndexReader.loadDataBlockWithScanInfo(HFileBlockIndex.java:253)
at
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:539)
at
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:560)
at
org.apache.hadoop.hbase.io.hfile.AbstractHFileReader$Scanner.seekTo(AbstractHFileReader.java:308)
at
org.apache.hadoop.hbase.io.HalfStoreFileReader$1.seekTo(HalfStoreFileReader.java:205)
at
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekAtOrAfter(StoreFileScanner.java:244)
at
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seek(StoreFileScanner.java:152)
... 12 more
Caused by: java.io.IOException: Invalid HFile block magic:
\x00\x00\x00\x00\x00\x00\x00\x00
at org.apache.hadoop.hbase.io.hfile.BlockType.parse(BlockType.java:154)
at org.apache.hadoop.hbase.io.hfile.BlockType.read(BlockType.java:165)
at
org.apache.hadoop.hbase.io.hfile.HFileBlock.<init>(HFileBlock.java:252)
at
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderV2.readBlockDataInternal(HFileBlock.java:1546)
... 21 more
was:
hi every body
my table has some cell that load with bulk load scenario and some cells for
increment.
we use 2 job to load data into table, first job use increment in reduce site
and second job use bulk load.
first we run increment job, next run bulk job, after that we got this exception:
2015-06-24 17:40:01,557 INFO
[regionserver60020-smallCompactions-1434448531302] regionserver.HRegion:
Starting compaction on c2 in region table1,\x04C#P1"\x07\x94
,1435065082383.0fe38a6c782600e4d46f1f148144b489.
2015-06-24 17:40:01,558 INFO
[regionserver60020-smallCompactions-1434448531302] regionserver.HStore:
Starting compaction of 3 file(s) in c2 of table1,\x04C#P1"\x07\x94
,1435065082383.0fe38a6c782600e4d46f1f148144b489. into
tmpdir=hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/.tmp,
totalSize=43.1m
2015-06-24 17:40:01,558 DEBUG
[regionserver60020-smallCompactions-1434448531302] regionserver.StoreFileInfo:
reference
'hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/6b1249a3b474474db5cf6c664f2d98dc.d21f8ee8b3c915fd9e1c143a0f1892e5'
to region=d21f8ee8b3c915fd9e1c143a0f1892e5
hfile=6b1249a3b474474db5cf6c664f2d98dc
2015-06-24 17:40:01,558 DEBUG
[regionserver60020-smallCompactions-1434448531302] compactions.Compactor:
Compacting
hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/6b1249a3b474474db5cf6c664f2d98dc.d21f8ee8b3c915fd9e1c143a0f1892e5-hdfs://m2/hbase2/data/default/table1/d21f8ee8b3c915fd9e1c143a0f1892e5/c2/6b1249a3b474474db5cf6c664f2d98dc-top,
keycount=575485, bloomtype=ROW, size=20.8m, encoding=NONE, seqNum=9,
earliestPutTs=1434875448405
2015-06-24 17:40:01,558 DEBUG
[regionserver60020-smallCompactions-1434448531302] compactions.Compactor:
Compacting
hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/41e13b20ee79435ebc260d11d3bf9920_SeqId_11_,
keycount=562988, bloomtype=ROW, size=10.1m, encoding=NONE, seqNum=11,
earliestPutTs=1435076732205
2015-06-24 17:40:01,558 DEBUG
[regionserver60020-smallCompactions-1434448531302] compactions.Compactor:
Compacting
hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/565c45ff05b14a419978834c86defa1a_SeqId_12_,
keycount=554577, bloomtype=ROW, size=12.2m, encoding=NONE, seqNum=12,
earliestPutTs=1435136926850
2015-06-24 17:40:01,560 ERROR
[regionserver60020-smallCompactions-1434448531302]
regionserver.CompactSplitThread: Compaction failed Request =
regionName=table1,\x04C#P1"\x07\x94
,1435065082383.0fe38a6c782600e4d46f1f148144b489., storeName=c2, fileCount=3,
fileSize=43.1m (20.8m, 10.1m, 12.2m), priority=1, time=6077271921381072
java.io.IOException: Could not seek
StoreFileScanner[org.apache.hadoop.hbase.io.HalfStoreFileReader$1@1d1eb574,
cur=null] to key /c2:/LATEST_TIMESTAMP/DeleteFamily/vlen=0/mvcc=0
at
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seek(StoreFileScanner.java:164)
at
org.apache.hadoop.hbase.regionserver.StoreScanner.seekScanners(StoreScanner.java:329)
at
org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:252)
at
org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:214)
at
org.apache.hadoop.hbase.regionserver.compactions.Compactor.createScanner(Compactor.java:299)
at
org.apache.hadoop.hbase.regionserver.compactions.DefaultCompactor.compact(DefaultCompactor.java:87)
at
org.apache.hadoop.hbase.regionserver.DefaultStoreEngine$DefaultCompactionContext.compact(DefaultStoreEngine.java:112)
at org.apache.hadoop.hbase.regionserver.HStore.compact(HStore.java:1113)
at
org.apache.hadoop.hbase.regionserver.HRegion.compact(HRegion.java:1519)
at
org.apache.hadoop.hbase.regionserver.CompactSplitThread$CompactionRunner.run(CompactSplitThread.java:498)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: Failed to read compressed block at 10930320,
onDiskSizeWithoutHeader=22342, preReadHeaderSize=33, header.length=33, header
bytes:
\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00
at
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderV2.readBlockDataInternal(HFileBlock.java:1549)
at
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderV2.readBlockData(HFileBlock.java:1413)
at
org.apache.hadoop.hbase.io.hfile.HFileReaderV2.readBlock(HFileReaderV2.java:394)
at
org.apache.hadoop.hbase.io.hfile.HFileBlockIndex$BlockIndexReader.loadDataBlockWithScanInfo(HFileBlockIndex.java:253)
at
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:539)
at
org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:560)
at
org.apache.hadoop.hbase.io.hfile.AbstractHFileReader$Scanner.seekTo(AbstractHFileReader.java:308)
at
org.apache.hadoop.hbase.io.HalfStoreFileReader$1.seekTo(HalfStoreFileReader.java:205)
at
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekAtOrAfter(StoreFileScanner.java:244)
at
org.apache.hadoop.hbase.regionserver.StoreFileScanner.seek(StoreFileScanner.java:152)
... 12 more
Caused by: java.io.IOException: Invalid HFile block magic:
\x00\x00\x00\x00\x00\x00\x00\x00
at org.apache.hadoop.hbase.io.hfile.BlockType.parse(BlockType.java:154)
at org.apache.hadoop.hbase.io.hfile.BlockType.read(BlockType.java:165)
at
org.apache.hadoop.hbase.io.hfile.HFileBlock.<init>(HFileBlock.java:252)
at
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderV2.readBlockDataInternal(HFileBlock.java:1546)
... 21 more
> Invalid HFile block magic
> -------------------------
>
> Key: HBASE-13962
> URL: https://issues.apache.org/jira/browse/HBASE-13962
> Project: HBase
> Issue Type: Bug
> Affects Versions: 0.98.12.1
> Environment: hadoop 1.2.1
> hbase 0.98.12.1
> jdk 1.7.0.79
> os : ubuntu 12.04.1 amd64
> Reporter: reaz hedayati
>
> hi every body
> my table has some cell that load with bulk load scenario and some cells for
> increment.
> we use 2 job to load data into table, first job use increment in reduce site
> and second job use bulk load.
> first we run increment job, next run bulk job and run completebulkload job,
> after that we got this exception:
> 2015-06-24 17:40:01,557 INFO
> [regionserver60020-smallCompactions-1434448531302] regionserver.HRegion:
> Starting compaction on c2 in region table1,\x04C#P1"\x07\x94
> ,1435065082383.0fe38a6c782600e4d46f1f148144b489.
> 2015-06-24 17:40:01,558 INFO
> [regionserver60020-smallCompactions-1434448531302] regionserver.HStore:
> Starting compaction of 3 file(s) in c2 of table1,\x04C#P1"\x07\x94
> ,1435065082383.0fe38a6c782600e4d46f1f148144b489. into
> tmpdir=hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/.tmp,
> totalSize=43.1m
> 2015-06-24 17:40:01,558 DEBUG
> [regionserver60020-smallCompactions-1434448531302]
> regionserver.StoreFileInfo: reference
> 'hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/6b1249a3b474474db5cf6c664f2d98dc.d21f8ee8b3c915fd9e1c143a0f1892e5'
> to region=d21f8ee8b3c915fd9e1c143a0f1892e5
> hfile=6b1249a3b474474db5cf6c664f2d98dc
> 2015-06-24 17:40:01,558 DEBUG
> [regionserver60020-smallCompactions-1434448531302] compactions.Compactor:
> Compacting
> hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/6b1249a3b474474db5cf6c664f2d98dc.d21f8ee8b3c915fd9e1c143a0f1892e5-hdfs://m2/hbase2/data/default/table1/d21f8ee8b3c915fd9e1c143a0f1892e5/c2/6b1249a3b474474db5cf6c664f2d98dc-top,
> keycount=575485, bloomtype=ROW, size=20.8m, encoding=NONE, seqNum=9,
> earliestPutTs=1434875448405
> 2015-06-24 17:40:01,558 DEBUG
> [regionserver60020-smallCompactions-1434448531302] compactions.Compactor:
> Compacting
> hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/41e13b20ee79435ebc260d11d3bf9920_SeqId_11_,
> keycount=562988, bloomtype=ROW, size=10.1m, encoding=NONE, seqNum=11,
> earliestPutTs=1435076732205
> 2015-06-24 17:40:01,558 DEBUG
> [regionserver60020-smallCompactions-1434448531302] compactions.Compactor:
> Compacting
> hdfs://m2/hbase2/data/default/table1/0fe38a6c782600e4d46f1f148144b489/c2/565c45ff05b14a419978834c86defa1a_SeqId_12_,
> keycount=554577, bloomtype=ROW, size=12.2m, encoding=NONE, seqNum=12,
> earliestPutTs=1435136926850
> 2015-06-24 17:40:01,560 ERROR
> [regionserver60020-smallCompactions-1434448531302]
> regionserver.CompactSplitThread: Compaction failed Request =
> regionName=table1,\x04C#P1"\x07\x94
> ,1435065082383.0fe38a6c782600e4d46f1f148144b489., storeName=c2, fileCount=3,
> fileSize=43.1m (20.8m, 10.1m, 12.2m), priority=1, time=6077271921381072
> java.io.IOException: Could not seek
> StoreFileScanner[org.apache.hadoop.hbase.io.HalfStoreFileReader$1@1d1eb574,
> cur=null] to key /c2:/LATEST_TIMESTAMP/DeleteFamily/vlen=0/mvcc=0
> at
> org.apache.hadoop.hbase.regionserver.StoreFileScanner.seek(StoreFileScanner.java:164)
> at
> org.apache.hadoop.hbase.regionserver.StoreScanner.seekScanners(StoreScanner.java:329)
> at
> org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:252)
> at
> org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:214)
> at
> org.apache.hadoop.hbase.regionserver.compactions.Compactor.createScanner(Compactor.java:299)
> at
> org.apache.hadoop.hbase.regionserver.compactions.DefaultCompactor.compact(DefaultCompactor.java:87)
> at
> org.apache.hadoop.hbase.regionserver.DefaultStoreEngine$DefaultCompactionContext.compact(DefaultStoreEngine.java:112)
> at
> org.apache.hadoop.hbase.regionserver.HStore.compact(HStore.java:1113)
> at
> org.apache.hadoop.hbase.regionserver.HRegion.compact(HRegion.java:1519)
> at
> org.apache.hadoop.hbase.regionserver.CompactSplitThread$CompactionRunner.run(CompactSplitThread.java:498)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.io.IOException: Failed to read compressed block at 10930320,
> onDiskSizeWithoutHeader=22342, preReadHeaderSize=33, header.length=33, header
> bytes:
> \x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00
> at
> org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderV2.readBlockDataInternal(HFileBlock.java:1549)
> at
> org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderV2.readBlockData(HFileBlock.java:1413)
> at
> org.apache.hadoop.hbase.io.hfile.HFileReaderV2.readBlock(HFileReaderV2.java:394)
> at
> org.apache.hadoop.hbase.io.hfile.HFileBlockIndex$BlockIndexReader.loadDataBlockWithScanInfo(HFileBlockIndex.java:253)
> at
> org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:539)
> at
> org.apache.hadoop.hbase.io.hfile.HFileReaderV2$AbstractScannerV2.seekTo(HFileReaderV2.java:560)
> at
> org.apache.hadoop.hbase.io.hfile.AbstractHFileReader$Scanner.seekTo(AbstractHFileReader.java:308)
> at
> org.apache.hadoop.hbase.io.HalfStoreFileReader$1.seekTo(HalfStoreFileReader.java:205)
> at
> org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekAtOrAfter(StoreFileScanner.java:244)
> at
> org.apache.hadoop.hbase.regionserver.StoreFileScanner.seek(StoreFileScanner.java:152)
> ... 12 more
> Caused by: java.io.IOException: Invalid HFile block magic:
> \x00\x00\x00\x00\x00\x00\x00\x00
> at
> org.apache.hadoop.hbase.io.hfile.BlockType.parse(BlockType.java:154)
> at org.apache.hadoop.hbase.io.hfile.BlockType.read(BlockType.java:165)
> at
> org.apache.hadoop.hbase.io.hfile.HFileBlock.<init>(HFileBlock.java:252)
> at
> org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderV2.readBlockDataInternal(HFileBlock.java:1546)
> ... 21 more
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)