HBASE-15296 Break out writer and reader from StoreFile. Done using Intellij15 Refactor > Move. (Apekshit)
Change-Id: Ie719569cc3393e0b5361e9d462c3cf125ad5144e Signed-off-by: stack <[email protected]> Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/7efb9ede Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/7efb9ede Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/7efb9ede Branch: refs/heads/master Commit: 7efb9edecbdf8b35046230575d504e4caeb80f34 Parents: dbdfd8e Author: Apekshit <[email protected]> Authored: Tue Apr 12 13:30:00 2016 -0700 Committer: stack <[email protected]> Committed: Wed Apr 13 22:43:03 2016 -0700 ---------------------------------------------------------------------- .../hbase/coprocessor/BaseRegionObserver.java | 10 +- .../hbase/coprocessor/RegionObserver.java | 9 +- .../hadoop/hbase/io/HalfStoreFileReader.java | 4 +- .../io/hfile/CompoundBloomFilterWriter.java | 3 +- .../hbase/mapreduce/HFileOutputFormat2.java | 9 +- .../hbase/mapreduce/LoadIncrementalHFiles.java | 6 +- .../hbase/mob/DefaultMobStoreCompactor.java | 26 +- .../hbase/mob/DefaultMobStoreFlusher.java | 8 +- .../org/apache/hadoop/hbase/mob/MobUtils.java | 15 +- .../compactions/PartitionedMobCompactor.java | 12 +- .../hbase/mob/mapreduce/MemStoreWrapper.java | 4 +- .../regionserver/AbstractMultiFileWriter.java | 13 +- .../regionserver/DateTieredMultiFileWriter.java | 13 +- .../hbase/regionserver/DefaultStoreFlusher.java | 2 +- .../hadoop/hbase/regionserver/HMobStore.java | 10 +- .../hadoop/hbase/regionserver/HRegion.java | 2 +- .../hadoop/hbase/regionserver/HStore.java | 30 +- .../regionserver/RegionCoprocessorHost.java | 10 +- .../apache/hadoop/hbase/regionserver/Store.java | 4 +- .../hadoop/hbase/regionserver/StoreFile.java | 1099 +----------------- .../hbase/regionserver/StoreFileInfo.java | 6 +- .../hbase/regionserver/StoreFileReader.java | 647 +++++++++++ .../hbase/regionserver/StoreFileScanner.java | 9 +- .../hbase/regionserver/StoreFileWriter.java | 514 ++++++++ .../hadoop/hbase/regionserver/StoreFlusher.java | 2 +- .../hadoop/hbase/regionserver/StoreUtils.java | 2 +- .../regionserver/StripeMultiFileWriter.java | 15 +- .../hbase/regionserver/StripeStoreFlusher.java | 6 +- .../AbstractMultiOutputCompactor.java | 4 +- .../compactions/CompactionRequest.java | 4 +- .../regionserver/compactions/Compactor.java | 8 +- .../compactions/DefaultCompactor.java | 26 +- .../hbase/coprocessor/SimpleRegionObserver.java | 10 +- .../hadoop/hbase/io/hfile/TestCacheOnWrite.java | 4 +- .../hadoop/hbase/io/hfile/TestPrefetch.java | 4 +- .../hfile/TestSeekBeforeWithInlineBlocks.java | 6 +- .../apache/hadoop/hbase/mob/MobTestUtil.java | 6 +- .../hadoop/hbase/mob/TestCachedMobFile.java | 10 +- .../apache/hadoop/hbase/mob/TestMobFile.java | 5 +- .../hadoop/hbase/mob/TestMobFileCache.java | 4 +- .../TestPartitionedMobCompactor.java | 6 +- .../regionserver/CreateRandomStoreFile.java | 2 +- .../regionserver/DataBlockEncodingTool.java | 2 +- .../EncodedSeekPerformanceTest.java | 4 +- .../hbase/regionserver/MockStoreFile.java | 5 +- .../regionserver/TestCacheOnWriteInSchema.java | 4 +- .../hbase/regionserver/TestCompaction.java | 2 +- .../regionserver/TestCompoundBloomFilter.java | 4 +- .../hbase/regionserver/TestFSErrorsExposed.java | 6 +- .../hbase/regionserver/TestHMobStore.java | 2 +- .../hadoop/hbase/regionserver/TestHRegion.java | 4 +- .../regionserver/TestReversibleScanners.java | 18 +- .../regionserver/TestScanWithBloomError.java | 2 +- .../hadoop/hbase/regionserver/TestStore.java | 6 +- .../hbase/regionserver/TestStoreFile.java | 60 +- .../TestStoreFileScannerWithTagCompression.java | 6 +- .../regionserver/TestStripeStoreEngine.java | 2 +- .../compactions/MockStoreFileGenerator.java | 3 +- .../regionserver/compactions/TestCompactor.java | 12 +- .../compactions/TestStripeCompactionPolicy.java | 3 +- .../hadoop/hbase/spark/HBaseContext.scala | 10 +- 61 files changed, 1412 insertions(+), 1332 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java index da7252b..67093f8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java @@ -55,7 +55,7 @@ import org.apache.hadoop.hbase.regionserver.RegionScanner; import org.apache.hadoop.hbase.regionserver.ScanType; import org.apache.hadoop.hbase.regionserver.Store; import org.apache.hadoop.hbase.regionserver.StoreFile; -import org.apache.hadoop.hbase.regionserver.StoreFile.Reader; +import org.apache.hadoop.hbase.regionserver.StoreFileReader; import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest; import org.apache.hadoop.hbase.regionserver.wal.HLogKey; import org.apache.hadoop.hbase.regionserver.wal.WALEdit; @@ -485,16 +485,16 @@ public class BaseRegionObserver implements RegionObserver { } @Override - public Reader preStoreFileReaderOpen(ObserverContext<RegionCoprocessorEnvironment> ctx, + public StoreFileReader preStoreFileReaderOpen(ObserverContext<RegionCoprocessorEnvironment> ctx, FileSystem fs, Path p, FSDataInputStreamWrapper in, long size, CacheConfig cacheConf, - Reference r, Reader reader) throws IOException { + Reference r, StoreFileReader reader) throws IOException { return reader; } @Override - public Reader postStoreFileReaderOpen(ObserverContext<RegionCoprocessorEnvironment> ctx, + public StoreFileReader postStoreFileReaderOpen(ObserverContext<RegionCoprocessorEnvironment> ctx, FileSystem fs, Path p, FSDataInputStreamWrapper in, long size, CacheConfig cacheConf, - Reference r, Reader reader) throws IOException { + Reference r, StoreFileReader reader) throws IOException { return reader; } http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java index e133f73..db08b76 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java @@ -55,6 +55,7 @@ import org.apache.hadoop.hbase.regionserver.RegionScanner; import org.apache.hadoop.hbase.regionserver.ScanType; import org.apache.hadoop.hbase.regionserver.Store; import org.apache.hadoop.hbase.regionserver.StoreFile; +import org.apache.hadoop.hbase.regionserver.StoreFileReader; import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest; import org.apache.hadoop.hbase.regionserver.wal.HLogKey; import org.apache.hadoop.hbase.wal.WALKey; @@ -1210,9 +1211,9 @@ public interface RegionObserver extends Coprocessor { * default behavior, null otherwise * @throws IOException */ - StoreFile.Reader preStoreFileReaderOpen(final ObserverContext<RegionCoprocessorEnvironment> ctx, + StoreFileReader preStoreFileReaderOpen(final ObserverContext<RegionCoprocessorEnvironment> ctx, final FileSystem fs, final Path p, final FSDataInputStreamWrapper in, long size, - final CacheConfig cacheConf, final Reference r, StoreFile.Reader reader) throws IOException; + final CacheConfig cacheConf, final Reference r, StoreFileReader reader) throws IOException; /** * Called after the creation of Reader for a store file. @@ -1228,9 +1229,9 @@ public interface RegionObserver extends Coprocessor { * @return The reader to use * @throws IOException */ - StoreFile.Reader postStoreFileReaderOpen(final ObserverContext<RegionCoprocessorEnvironment> ctx, + StoreFileReader postStoreFileReaderOpen(final ObserverContext<RegionCoprocessorEnvironment> ctx, final FileSystem fs, final Path p, final FSDataInputStreamWrapper in, long size, - final CacheConfig cacheConf, final Reference r, StoreFile.Reader reader) throws IOException; + final CacheConfig cacheConf, final Reference r, StoreFileReader reader) throws IOException; /** * Called after a new cell has been created during an increment operation, but before http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java index 067d24c..a4a281e 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java @@ -33,7 +33,7 @@ import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.io.hfile.CacheConfig; import org.apache.hadoop.hbase.io.hfile.HFileScanner; -import org.apache.hadoop.hbase.regionserver.StoreFile; +import org.apache.hadoop.hbase.regionserver.StoreFileReader; import org.apache.hadoop.hbase.util.Bytes; /** @@ -50,7 +50,7 @@ import org.apache.hadoop.hbase.util.Bytes; * <p>This file is not splitable. Calls to {@link #midkey()} return null. */ @InterfaceAudience.Private -public class HalfStoreFileReader extends StoreFile.Reader { +public class HalfStoreFileReader extends StoreFileReader { private static final Log LOG = LogFactory.getLog(HalfStoreFileReader.class); final boolean top; // This is the key we split around. Its the first possible entry on a row: http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java index cd60562..d95aafd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java @@ -29,6 +29,7 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.CellComparator; +import org.apache.hadoop.hbase.regionserver.StoreFileWriter; import org.apache.hadoop.hbase.util.BloomFilterChunk; import org.apache.hadoop.hbase.util.BloomFilterUtil; import org.apache.hadoop.hbase.util.BloomFilterWriter; @@ -155,7 +156,7 @@ public class CompoundBloomFilterWriter extends CompoundBloomFilterBase * Adds a Bloom filter key. This key must be greater than the previous key, * as defined by the comparator this compound Bloom filter is configured * with. For efficiency, key monotonicity is not checked here. See - * {@link org.apache.hadoop.hbase.regionserver.StoreFile.Writer#append( + * {@link StoreFileWriter#append( * org.apache.hadoop.hbase.Cell)} for the details of deduplication. */ @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java index 059e030..606d336 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java @@ -65,6 +65,7 @@ import org.apache.hadoop.hbase.io.hfile.HFileContextBuilder; import org.apache.hadoop.hbase.regionserver.BloomType; import org.apache.hadoop.hbase.regionserver.HStore; import org.apache.hadoop.hbase.regionserver.StoreFile; +import org.apache.hadoop.hbase.regionserver.StoreFileWriter; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.NullWritable; import org.apache.hadoop.io.SequenceFile; @@ -305,12 +306,12 @@ public class HFileOutputFormat2 if (null == favoredNodes) { wl.writer = - new StoreFile.WriterBuilder(conf, new CacheConfig(tempConf), fs) + new StoreFileWriter.Builder(conf, new CacheConfig(tempConf), fs) .withOutputDir(familydir).withBloomType(bloomType) .withComparator(CellComparator.COMPARATOR).withFileContext(hFileContext).build(); } else { wl.writer = - new StoreFile.WriterBuilder(conf, new CacheConfig(tempConf), new HFileSystem(fs)) + new StoreFileWriter.Builder(conf, new CacheConfig(tempConf), new HFileSystem(fs)) .withOutputDir(familydir).withBloomType(bloomType) .withComparator(CellComparator.COMPARATOR).withFileContext(hFileContext) .withFavoredNodes(favoredNodes).build(); @@ -320,7 +321,7 @@ public class HFileOutputFormat2 return wl; } - private void close(final StoreFile.Writer w) throws IOException { + private void close(final StoreFileWriter w) throws IOException { if (w != null) { w.appendFileInfo(StoreFile.BULKLOAD_TIME_KEY, Bytes.toBytes(System.currentTimeMillis())); @@ -350,7 +351,7 @@ public class HFileOutputFormat2 */ static class WriterLength { long written = 0; - StoreFile.Writer writer = null; + StoreFileWriter writer = null; } /** http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java index 5d75d56..86a84a4 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java @@ -65,8 +65,8 @@ import org.apache.hadoop.hbase.io.hfile.HFileScanner; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.regionserver.BloomType; import org.apache.hadoop.hbase.regionserver.HStore; -import org.apache.hadoop.hbase.regionserver.StoreFile; import org.apache.hadoop.hbase.regionserver.StoreFileInfo; +import org.apache.hadoop.hbase.regionserver.StoreFileWriter; import org.apache.hadoop.hbase.security.UserProvider; import org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint; import org.apache.hadoop.hbase.security.token.FsDelegationToken; @@ -890,7 +890,7 @@ public class LoadIncrementalHFiles extends Configured implements Tool { FileSystem fs = inFile.getFileSystem(conf); CacheConfig cacheConf = new CacheConfig(conf); HalfStoreFileReader halfReader = null; - StoreFile.Writer halfWriter = null; + StoreFileWriter halfWriter = null; try { halfReader = new HalfStoreFileReader(fs, inFile, cacheConf, reference, conf); Map<byte[], byte[]> fileInfo = halfReader.loadFileInfo(); @@ -906,7 +906,7 @@ public class LoadIncrementalHFiles extends Configured implements Tool { .withDataBlockEncoding(familyDescriptor.getDataBlockEncoding()) .withIncludesTags(true) .build(); - halfWriter = new StoreFile.WriterBuilder(conf, cacheConf, + halfWriter = new StoreFileWriter.Builder(conf, cacheConf, fs) .withFilePath(outFile) .withBloomType(bloomFilterType) http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java index fe640c6..711f31d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java @@ -42,8 +42,8 @@ import org.apache.hadoop.hbase.regionserver.MobCompactionStoreScanner; import org.apache.hadoop.hbase.regionserver.ScanType; import org.apache.hadoop.hbase.regionserver.ScannerContext; import org.apache.hadoop.hbase.regionserver.Store; -import org.apache.hadoop.hbase.regionserver.StoreFile.Writer; import org.apache.hadoop.hbase.regionserver.StoreFileScanner; +import org.apache.hadoop.hbase.regionserver.StoreFileWriter; import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest; import org.apache.hadoop.hbase.regionserver.compactions.DefaultCompactor; import org.apache.hadoop.hbase.regionserver.throttle.ThroughputController; @@ -86,17 +86,17 @@ public class DefaultMobStoreCompactor extends DefaultCompactor { } }; - private final CellSinkFactory<Writer> writerFactory = new CellSinkFactory<Writer>() { - - @Override - public Writer createWriter(InternalScanner scanner, - org.apache.hadoop.hbase.regionserver.compactions.Compactor.FileDetails fd, - boolean shouldDropBehind) throws IOException { - // make this writer with tags always because of possible new cells with tags. - return store.createWriterInTmp(fd.maxKeyCount, compactionCompression, true, true, true, - shouldDropBehind); - } - }; + private final CellSinkFactory<StoreFileWriter> writerFactory = + new CellSinkFactory<StoreFileWriter>() { + @Override + public StoreFileWriter createWriter(InternalScanner scanner, + org.apache.hadoop.hbase.regionserver.compactions.Compactor.FileDetails fd, + boolean shouldDropBehind) throws IOException { + // make this writer with tags always because of possible new cells with tags. + return store.createWriterInTmp(fd.maxKeyCount, compactionCompression, true, true, true, + shouldDropBehind); + } + }; public DefaultMobStoreCompactor(Configuration conf, Store store) { super(conf, store); @@ -180,7 +180,7 @@ public class DefaultMobStoreCompactor extends DefaultCompactor { boolean hasMore; Path path = MobUtils.getMobFamilyPath(conf, store.getTableName(), store.getColumnFamilyName()); byte[] fileName = null; - Writer mobFileWriter = null, delFileWriter = null; + StoreFileWriter mobFileWriter = null, delFileWriter = null; long mobCells = 0, deleteMarkersCount = 0; Tag tableNameTag = new ArrayBackedTag(TagType.MOB_TABLE_NAME_TAG_TYPE, store.getTableName().getName()); http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java index eb6c739..04a8782 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java @@ -41,7 +41,7 @@ import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.regionserver.MemStoreSnapshot; import org.apache.hadoop.hbase.regionserver.ScannerContext; import org.apache.hadoop.hbase.regionserver.Store; -import org.apache.hadoop.hbase.regionserver.StoreFile; +import org.apache.hadoop.hbase.regionserver.StoreFileWriter; import org.apache.hadoop.hbase.regionserver.throttle.ThroughputController; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.util.StringUtils; @@ -108,7 +108,7 @@ public class DefaultMobStoreFlusher extends DefaultStoreFlusher { if (scanner == null) { return result; // NULL scanner returned from coprocessor hooks means skip normal processing } - StoreFile.Writer writer; + StoreFileWriter writer; try { // TODO: We can fail in the below block before we complete adding this flush to // list of store files. Add cleanup of anything put on filesystem if we fail. @@ -155,8 +155,8 @@ public class DefaultMobStoreFlusher extends DefaultStoreFlusher { * @throws IOException */ protected void performMobFlush(MemStoreSnapshot snapshot, long cacheFlushId, - InternalScanner scanner, StoreFile.Writer writer, MonitoredTask status) throws IOException { - StoreFile.Writer mobFileWriter = null; + InternalScanner scanner, StoreFileWriter writer, MonitoredTask status) throws IOException { + StoreFileWriter mobFileWriter = null; int compactionKVMax = conf.getInt(HConstants.COMPACTION_KV_MAX, HConstants.COMPACTION_KV_MAX_DEFAULT); long mobCount = 0; http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java index 52a19f5..ecd2415 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java @@ -68,6 +68,7 @@ import org.apache.hadoop.hbase.mob.compactions.PartitionedMobCompactor; import org.apache.hadoop.hbase.regionserver.BloomType; import org.apache.hadoop.hbase.regionserver.HStore; import org.apache.hadoop.hbase.regionserver.StoreFile; +import org.apache.hadoop.hbase.regionserver.StoreFileWriter; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import org.apache.hadoop.hbase.util.FSUtils; @@ -462,7 +463,7 @@ public final class MobUtils { * @return The writer for the mob file. * @throws IOException */ - public static StoreFile.Writer createWriter(Configuration conf, FileSystem fs, + public static StoreFileWriter createWriter(Configuration conf, FileSystem fs, HColumnDescriptor family, String date, Path basePath, long maxKeyCount, Compression.Algorithm compression, String startKey, CacheConfig cacheConfig, Encryption.Context cryptoContext) @@ -485,7 +486,7 @@ public final class MobUtils { * @return The writer for the mob file. * @throws IOException */ - public static StoreFile.Writer createRefFileWriter(Configuration conf, FileSystem fs, + public static StoreFileWriter createRefFileWriter(Configuration conf, FileSystem fs, HColumnDescriptor family, Path basePath, long maxKeyCount, CacheConfig cacheConfig, Encryption.Context cryptoContext) throws IOException { @@ -497,7 +498,7 @@ public final class MobUtils { .withEncryptionContext(cryptoContext).withCreateTime(EnvironmentEdgeManager.currentTime()) .build(); Path tempPath = new Path(basePath, UUID.randomUUID().toString().replaceAll("-", "")); - StoreFile.Writer w = new StoreFile.WriterBuilder(conf, cacheConfig, fs).withFilePath(tempPath) + StoreFileWriter w = new StoreFileWriter.Builder(conf, cacheConfig, fs).withFilePath(tempPath) .withComparator(CellComparator.COMPARATOR).withBloomType(family.getBloomFilterType()) .withMaxKeyCount(maxKeyCount).withFileContext(hFileContext).build(); return w; @@ -518,7 +519,7 @@ public final class MobUtils { * @return The writer for the mob file. * @throws IOException */ - public static StoreFile.Writer createWriter(Configuration conf, FileSystem fs, + public static StoreFileWriter createWriter(Configuration conf, FileSystem fs, HColumnDescriptor family, String date, Path basePath, long maxKeyCount, Compression.Algorithm compression, byte[] startKey, CacheConfig cacheConfig, Encryption.Context cryptoContext) @@ -544,7 +545,7 @@ public final class MobUtils { * @return The writer for the del file. * @throws IOException */ - public static StoreFile.Writer createDelFileWriter(Configuration conf, FileSystem fs, + public static StoreFileWriter createDelFileWriter(Configuration conf, FileSystem fs, HColumnDescriptor family, String date, Path basePath, long maxKeyCount, Compression.Algorithm compression, byte[] startKey, CacheConfig cacheConfig, Encryption.Context cryptoContext) @@ -570,7 +571,7 @@ public final class MobUtils { * @return The writer for the mob file. * @throws IOException */ - private static StoreFile.Writer createWriter(Configuration conf, FileSystem fs, + private static StoreFileWriter createWriter(Configuration conf, FileSystem fs, HColumnDescriptor family, MobFileName mobFileName, Path basePath, long maxKeyCount, Compression.Algorithm compression, CacheConfig cacheConfig, Encryption.Context cryptoContext) throws IOException { @@ -583,7 +584,7 @@ public final class MobUtils { .withEncryptionContext(cryptoContext) .withCreateTime(EnvironmentEdgeManager.currentTime()).build(); - StoreFile.Writer w = new StoreFile.WriterBuilder(conf, cacheConfig, fs) + StoreFileWriter w = new StoreFileWriter.Builder(conf, cacheConfig, fs) .withFilePath(new Path(basePath, mobFileName.getFileName())) .withComparator(CellComparator.COMPARATOR).withBloomType(BloomType.NONE) .withMaxKeyCount(maxKeyCount).withFileContext(hFileContext).build(); http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java index 6c6f115..cfe76ae 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java @@ -69,9 +69,9 @@ import org.apache.hadoop.hbase.regionserver.ScanInfo; import org.apache.hadoop.hbase.regionserver.ScanType; import org.apache.hadoop.hbase.regionserver.ScannerContext; import org.apache.hadoop.hbase.regionserver.StoreFile; -import org.apache.hadoop.hbase.regionserver.StoreFile.Writer; import org.apache.hadoop.hbase.regionserver.StoreFileInfo; import org.apache.hadoop.hbase.regionserver.StoreFileScanner; +import org.apache.hadoop.hbase.regionserver.StoreFileWriter; import org.apache.hadoop.hbase.regionserver.StoreScanner; import org.apache.hadoop.hbase.security.EncryptionUtil; import org.apache.hadoop.hbase.util.Bytes; @@ -385,8 +385,8 @@ public class PartitionedMobCompactor extends MobCompactor { // Pair(maxSeqId, cellsCount) Pair<Long, Long> fileInfo = getFileInfo(mobFilesToCompact); // open writers for the mob files and new ref store files. - Writer writer = null; - Writer refFileWriter = null; + StoreFileWriter writer = null; + StoreFileWriter refFileWriter = null; Path filePath = null; Path refFilePath = null; long mobCells = 0; @@ -499,7 +499,7 @@ public class PartitionedMobCompactor extends MobCompactor { List<StoreFile> delFiles) throws IOException { // create a scanner for the del files. StoreScanner scanner = createScanner(delFiles, ScanType.COMPACT_RETAIN_DELETES); - Writer writer = null; + StoreFileWriter writer = null; Path filePath = null; try { writer = MobUtils.createDelFileWriter(conf, fs, column, @@ -589,7 +589,7 @@ public class PartitionedMobCompactor extends MobCompactor { * @param mobCellsCount The number of mob cells. * @throws IOException */ - private void closeMobFileWriter(Writer writer, long maxSeqId, long mobCellsCount) + private void closeMobFileWriter(StoreFileWriter writer, long maxSeqId, long mobCellsCount) throws IOException { if (writer != null) { writer.appendMetadata(maxSeqId, false, mobCellsCount); @@ -608,7 +608,7 @@ public class PartitionedMobCompactor extends MobCompactor { * @param bulkloadTime The timestamp at which the bulk load file is created. * @throws IOException */ - private void closeRefFileWriter(Writer writer, long maxSeqId, long bulkloadTime) + private void closeRefFileWriter(StoreFileWriter writer, long maxSeqId, long bulkloadTime) throws IOException { if (writer != null) { writer.appendMetadata(maxSeqId, false); http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java index 5955cc2..7997b49 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java @@ -45,7 +45,7 @@ import org.apache.hadoop.hbase.mob.mapreduce.SweepJob.SweepCounter; import org.apache.hadoop.hbase.regionserver.KeyValueScanner; import org.apache.hadoop.hbase.regionserver.MemStore; import org.apache.hadoop.hbase.regionserver.MemStoreSnapshot; -import org.apache.hadoop.hbase.regionserver.StoreFile; +import org.apache.hadoop.hbase.regionserver.StoreFileWriter; import org.apache.hadoop.hbase.security.EncryptionUtil; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.mapreduce.Reducer.Context; @@ -132,7 +132,7 @@ public class MemStoreWrapper { } // generate the files into a temp directory. String tempPathString = context.getConfiguration().get(SweepJob.WORKING_FILES_DIR_KEY); - StoreFile.Writer mobFileWriter = MobUtils.createWriter(conf, fs, hcd, partitionId.getDate(), + StoreFileWriter mobFileWriter = MobUtils.createWriter(conf, fs, hcd, partitionId.getDate(), new Path(tempPathString), snapshot.getCellsCount(), hcd.getCompactionCompression(), partitionId.getStartKey(), cacheConfig, cryptoContext); http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java index 4987c59..a4e0285 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java @@ -26,7 +26,6 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.regionserver.StoreFile.Writer; import org.apache.hadoop.hbase.regionserver.compactions.Compactor.CellSink; /** @@ -44,7 +43,7 @@ public abstract class AbstractMultiFileWriter implements CellSink { protected StoreScanner sourceScanner; public interface WriterFactory { - public StoreFile.Writer createWriter() throws IOException; + public StoreFileWriter createWriter() throws IOException; } /** @@ -66,13 +65,13 @@ public abstract class AbstractMultiFileWriter implements CellSink { */ public List<Path> commitWriters(long maxSeqId, boolean majorCompaction) throws IOException { preCommitWriters(); - Collection<StoreFile.Writer> writers = this.writers(); + Collection<StoreFileWriter> writers = this.writers(); if (LOG.isDebugEnabled()) { LOG.debug("Commit " + writers.size() + " writers, maxSeqId=" + maxSeqId + ", majorCompaction=" + majorCompaction); } List<Path> paths = new ArrayList<Path>(); - for (Writer writer : writers) { + for (StoreFileWriter writer : writers) { if (writer == null) { continue; } @@ -89,7 +88,7 @@ public abstract class AbstractMultiFileWriter implements CellSink { */ public List<Path> abortWriters() { List<Path> paths = new ArrayList<Path>(); - for (StoreFile.Writer writer : writers()) { + for (StoreFileWriter writer : writers()) { try { if (writer != null) { paths.add(writer.getPath()); @@ -102,7 +101,7 @@ public abstract class AbstractMultiFileWriter implements CellSink { return paths; } - protected abstract Collection<StoreFile.Writer> writers(); + protected abstract Collection<StoreFileWriter> writers(); /** * Subclasses override this method to be called at the end of a successful sequence of append; all @@ -115,6 +114,6 @@ public abstract class AbstractMultiFileWriter implements CellSink { * Subclasses override this method to be called before we close the give writer. Usually you can * append extra metadata to the writer. */ - protected void preCloseWriter(StoreFile.Writer writer) throws IOException { + protected void preCloseWriter(StoreFileWriter writer) throws IOException { } } http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java index f0bd444..2cea92f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java @@ -26,7 +26,6 @@ import java.util.TreeMap; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.regionserver.StoreFile.Writer; /** * class for cell sink that separates the provided cells into multiple files for date tiered @@ -35,8 +34,8 @@ import org.apache.hadoop.hbase.regionserver.StoreFile.Writer; @InterfaceAudience.Private public class DateTieredMultiFileWriter extends AbstractMultiFileWriter { - private final NavigableMap<Long, StoreFile.Writer> lowerBoundary2Writer - = new TreeMap<Long, StoreFile.Writer>(); + private final NavigableMap<Long, StoreFileWriter> lowerBoundary2Writer + = new TreeMap<Long, StoreFileWriter>(); private final boolean needEmptyFile; @@ -53,8 +52,8 @@ public class DateTieredMultiFileWriter extends AbstractMultiFileWriter { @Override public void append(Cell cell) throws IOException { - Map.Entry<Long, StoreFile.Writer> entry = lowerBoundary2Writer.floorEntry(cell.getTimestamp()); - StoreFile.Writer writer = entry.getValue(); + Map.Entry<Long, StoreFileWriter> entry = lowerBoundary2Writer.floorEntry(cell.getTimestamp()); + StoreFileWriter writer = entry.getValue(); if (writer == null) { writer = writerFactory.createWriter(); lowerBoundary2Writer.put(entry.getKey(), writer); @@ -63,7 +62,7 @@ public class DateTieredMultiFileWriter extends AbstractMultiFileWriter { } @Override - protected Collection<Writer> writers() { + protected Collection<StoreFileWriter> writers() { return lowerBoundary2Writer.values(); } @@ -72,7 +71,7 @@ public class DateTieredMultiFileWriter extends AbstractMultiFileWriter { if (!needEmptyFile) { return; } - for (StoreFile.Writer writer : lowerBoundary2Writer.values()) { + for (StoreFileWriter writer : lowerBoundary2Writer.values()) { if (writer != null) { return; } http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java index 711b987..bf8794a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java @@ -57,7 +57,7 @@ public class DefaultStoreFlusher extends StoreFlusher { return result; // NULL scanner returned from coprocessor hooks means skip normal processing } - StoreFile.Writer writer; + StoreFileWriter writer; try { // TODO: We can fail in the below block before we complete adding this flush to // list of store files. Add cleanup of anything put on filesystem if we fail. http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java index 7b44338..54b942d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java @@ -188,7 +188,7 @@ public class HMobStore extends HStore { * @return The writer for the mob file. * @throws IOException */ - public StoreFile.Writer createWriterInTmp(Date date, long maxKeyCount, + public StoreFileWriter createWriterInTmp(Date date, long maxKeyCount, Compression.Algorithm compression, byte[] startKey) throws IOException { if (startKey == null) { startKey = HConstants.EMPTY_START_ROW; @@ -208,7 +208,7 @@ public class HMobStore extends HStore { * @return The writer for the del file. * @throws IOException */ - public StoreFile.Writer createDelFileWriterInTmp(Date date, long maxKeyCount, + public StoreFileWriter createDelFileWriterInTmp(Date date, long maxKeyCount, Compression.Algorithm compression, byte[] startKey) throws IOException { if (startKey == null) { startKey = HConstants.EMPTY_START_ROW; @@ -230,7 +230,7 @@ public class HMobStore extends HStore { * @return The writer for the mob file. * @throws IOException */ - public StoreFile.Writer createWriterInTmp(String date, Path basePath, long maxKeyCount, + public StoreFileWriter createWriterInTmp(String date, Path basePath, long maxKeyCount, Compression.Algorithm compression, byte[] startKey) throws IOException { MobFileName mobFileName = MobFileName.create(startKey, date, UUID.randomUUID() .toString().replaceAll("-", "")); @@ -246,7 +246,7 @@ public class HMobStore extends HStore { * @return The writer for the mob file. * @throws IOException */ - public StoreFile.Writer createWriterInTmp(MobFileName mobFileName, Path basePath, + public StoreFileWriter createWriterInTmp(MobFileName mobFileName, Path basePath, long maxKeyCount, Compression.Algorithm compression) throws IOException { final CacheConfig writerCacheConf = mobCacheConfig; HFileContext hFileContext = new HFileContextBuilder().withCompression(compression) @@ -259,7 +259,7 @@ public class HMobStore extends HStore { .withEncryptionContext(cryptoContext) .withCreateTime(EnvironmentEdgeManager.currentTime()).build(); - StoreFile.Writer w = new StoreFile.WriterBuilder(conf, writerCacheConf, region.getFilesystem()) + StoreFileWriter w = new StoreFileWriter.Builder(conf, writerCacheConf, region.getFilesystem()) .withFilePath(new Path(basePath, mobFileName.getFileName())) .withComparator(CellComparator.COMPARATOR).withBloomType(BloomType.NONE) .withMaxKeyCount(maxKeyCount).withFileContext(hFileContext).build(); http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java index 5524de0..72238cc 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java @@ -1691,7 +1691,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi Collection<StoreFile> storeFiles = store.getStorefiles(); if (storeFiles == null) continue; for (StoreFile file : storeFiles) { - StoreFile.Reader sfReader = file.getReader(); + StoreFileReader sfReader = file.getReader(); if (sfReader == null) continue; HFile.Reader reader = sfReader.getHFileReader(); if (reader == null) continue; http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java index f7ad4ae..7468be0 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java @@ -611,7 +611,7 @@ public class HStore implements Store { info.setRegionCoprocessorHost(this.region.getCoprocessorHost()); StoreFile storeFile = new StoreFile(this.getFileSystem(), info, this.conf, this.cacheConf, this.family.getBloomFilterType()); - StoreFile.Reader r = storeFile.createReader(); + StoreFileReader r = storeFile.createReader(); r.setReplicaStoreFile(isPrimaryReplicaStore()); return storeFile; } @@ -749,7 +749,7 @@ public class HStore implements Store { } private void bulkLoadHFile(StoreFile sf) throws IOException { - StoreFile.Reader r = sf.getReader(); + StoreFileReader r = sf.getReader(); this.storeSize += r.length(); this.totalUncompressedBytes += r.getTotalUncompressedBytes(); @@ -917,7 +917,7 @@ public class HStore implements Store { status.setStatus("Flushing " + this + ": reopening flushed file"); StoreFile sf = createStoreFileAndReader(dstPath); - StoreFile.Reader r = sf.getReader(); + StoreFileReader r = sf.getReader(); this.storeSize += r.length(); this.totalUncompressedBytes += r.getTotalUncompressedBytes(); @@ -930,7 +930,7 @@ public class HStore implements Store { } @Override - public StoreFile.Writer createWriterInTmp(long maxKeyCount, Compression.Algorithm compression, + public StoreFileWriter createWriterInTmp(long maxKeyCount, Compression.Algorithm compression, boolean isCompaction, boolean includeMVCCReadpoint, boolean includesTag) throws IOException { @@ -947,7 +947,7 @@ public class HStore implements Store { * @return Writer for a new StoreFile in the tmp dir. */ @Override - public StoreFile.Writer createWriterInTmp(long maxKeyCount, Compression.Algorithm compression, + public StoreFileWriter createWriterInTmp(long maxKeyCount, Compression.Algorithm compression, boolean isCompaction, boolean includeMVCCReadpoint, boolean includesTag, boolean shouldDropBehind) throws IOException { @@ -966,7 +966,7 @@ public class HStore implements Store { } HFileContext hFileContext = createFileContext(compression, includeMVCCReadpoint, includesTag, cryptoContext); - StoreFile.Writer w = new StoreFile.WriterBuilder(conf, writerCacheConf, + StoreFileWriter w = new StoreFileWriter.Builder(conf, writerCacheConf, this.getFileSystem()) .withFilePath(fs.createTempName()) .withComparator(comparator) @@ -1749,7 +1749,7 @@ public class HStore implements Store { this.storeSize = 0L; this.totalUncompressedBytes = 0L; for (StoreFile hsf : this.storeEngine.getStoreFileManager().getStorefiles()) { - StoreFile.Reader r = hsf.getReader(); + StoreFileReader r = hsf.getReader(); if (r == null) { LOG.warn("StoreFile " + hsf + " has a null Reader"); continue; @@ -1900,7 +1900,7 @@ public class HStore implements Store { public long getMaxStoreFileAge() { long earliestTS = Long.MAX_VALUE; for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) { - StoreFile.Reader r = s.getReader(); + StoreFileReader r = s.getReader(); if (r == null) { LOG.warn("StoreFile " + s + " has a null Reader"); continue; @@ -1919,7 +1919,7 @@ public class HStore implements Store { public long getMinStoreFileAge() { long latestTS = 0; for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) { - StoreFile.Reader r = s.getReader(); + StoreFileReader r = s.getReader(); if (r == null) { LOG.warn("StoreFile " + s + " has a null Reader"); continue; @@ -1938,7 +1938,7 @@ public class HStore implements Store { public long getAvgStoreFileAge() { long sum = 0, count = 0; for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) { - StoreFile.Reader r = s.getReader(); + StoreFileReader r = s.getReader(); if (r == null) { LOG.warn("StoreFile " + s + " has a null Reader"); continue; @@ -1988,7 +1988,7 @@ public class HStore implements Store { public long getStorefilesSize() { long size = 0; for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) { - StoreFile.Reader r = s.getReader(); + StoreFileReader r = s.getReader(); if (r == null) { LOG.warn("StoreFile " + s + " has a null Reader"); continue; @@ -2002,7 +2002,7 @@ public class HStore implements Store { public long getStorefilesIndexSize() { long size = 0; for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) { - StoreFile.Reader r = s.getReader(); + StoreFileReader r = s.getReader(); if (r == null) { LOG.warn("StoreFile " + s + " has a null Reader"); continue; @@ -2016,7 +2016,7 @@ public class HStore implements Store { public long getTotalStaticIndexSize() { long size = 0; for (StoreFile s : this.storeEngine.getStoreFileManager().getStorefiles()) { - StoreFile.Reader r = s.getReader(); + StoreFileReader r = s.getReader(); if (r == null) { continue; } @@ -2029,7 +2029,7 @@ public class HStore implements Store { public long getTotalStaticBloomSize() { long size = 0; for (StoreFile s : this.storeEngine.getStoreFileManager().getStorefiles()) { - StoreFile.Reader r = s.getReader(); + StoreFileReader r = s.getReader(); if (r == null) { continue; } @@ -2430,7 +2430,7 @@ public class HStore implements Store { for (final StoreFile file : compactedfiles) { synchronized (file) { try { - StoreFile.Reader r = file.getReader(); + StoreFileReader r = file.getReader(); if (r == null) { if (LOG.isDebugEnabled()) { LOG.debug("The file " + file + " was closed but still not archived."); http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java index 48e67f7..6614b3e 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java @@ -1541,11 +1541,11 @@ public class RegionCoprocessorHost * default behavior, null otherwise * @throws IOException */ - public StoreFile.Reader preStoreFileReaderOpen(final FileSystem fs, final Path p, + public StoreFileReader preStoreFileReaderOpen(final FileSystem fs, final Path p, final FSDataInputStreamWrapper in, final long size, final CacheConfig cacheConf, final Reference r) throws IOException { return execOperationWithResult(null, - coprocessors.isEmpty() ? null : new RegionOperationWithResult<StoreFile.Reader>() { + coprocessors.isEmpty() ? null : new RegionOperationWithResult<StoreFileReader>() { @Override public void call(RegionObserver oserver, ObserverContext<RegionCoprocessorEnvironment> ctx) throws IOException { @@ -1565,11 +1565,11 @@ public class RegionCoprocessorHost * @return The reader to use * @throws IOException */ - public StoreFile.Reader postStoreFileReaderOpen(final FileSystem fs, final Path p, + public StoreFileReader postStoreFileReaderOpen(final FileSystem fs, final Path p, final FSDataInputStreamWrapper in, final long size, final CacheConfig cacheConf, - final Reference r, final StoreFile.Reader reader) throws IOException { + final Reference r, final StoreFileReader reader) throws IOException { return execOperationWithResult(reader, - coprocessors.isEmpty() ? null : new RegionOperationWithResult<StoreFile.Reader>() { + coprocessors.isEmpty() ? null : new RegionOperationWithResult<StoreFileReader>() { @Override public void call(RegionObserver oserver, ObserverContext<RegionCoprocessorEnvironment> ctx) throws IOException { http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java index 4cca4a1..b77a33b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java @@ -164,7 +164,7 @@ public interface Store extends HeapSize, StoreConfigInformation, PropagatingConf * @param includeMVCCReadpoint whether we should out the MVCC readpoint * @return Writer for a new StoreFile in the tmp dir. */ - StoreFile.Writer createWriterInTmp( + StoreFileWriter createWriterInTmp( long maxKeyCount, Compression.Algorithm compression, boolean isCompaction, @@ -180,7 +180,7 @@ public interface Store extends HeapSize, StoreConfigInformation, PropagatingConf * @param shouldDropBehind should the writer drop caches behind writes * @return Writer for a new StoreFile in the tmp dir. */ - StoreFile.Writer createWriterInTmp( + StoreFileWriter createWriterInTmp( long maxKeyCount, Compression.Algorithm compression, boolean isCompaction,
