HBASE-15296 Break out writer and reader from StoreFile. Done using Intellij15 
Refactor > Move. (Apekshit)

Change-Id: Ie719569cc3393e0b5361e9d462c3cf125ad5144e

Signed-off-by: stack <[email protected]>


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/7efb9ede
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/7efb9ede
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/7efb9ede

Branch: refs/heads/master
Commit: 7efb9edecbdf8b35046230575d504e4caeb80f34
Parents: dbdfd8e
Author: Apekshit <[email protected]>
Authored: Tue Apr 12 13:30:00 2016 -0700
Committer: stack <[email protected]>
Committed: Wed Apr 13 22:43:03 2016 -0700

----------------------------------------------------------------------
 .../hbase/coprocessor/BaseRegionObserver.java   |   10 +-
 .../hbase/coprocessor/RegionObserver.java       |    9 +-
 .../hadoop/hbase/io/HalfStoreFileReader.java    |    4 +-
 .../io/hfile/CompoundBloomFilterWriter.java     |    3 +-
 .../hbase/mapreduce/HFileOutputFormat2.java     |    9 +-
 .../hbase/mapreduce/LoadIncrementalHFiles.java  |    6 +-
 .../hbase/mob/DefaultMobStoreCompactor.java     |   26 +-
 .../hbase/mob/DefaultMobStoreFlusher.java       |    8 +-
 .../org/apache/hadoop/hbase/mob/MobUtils.java   |   15 +-
 .../compactions/PartitionedMobCompactor.java    |   12 +-
 .../hbase/mob/mapreduce/MemStoreWrapper.java    |    4 +-
 .../regionserver/AbstractMultiFileWriter.java   |   13 +-
 .../regionserver/DateTieredMultiFileWriter.java |   13 +-
 .../hbase/regionserver/DefaultStoreFlusher.java |    2 +-
 .../hadoop/hbase/regionserver/HMobStore.java    |   10 +-
 .../hadoop/hbase/regionserver/HRegion.java      |    2 +-
 .../hadoop/hbase/regionserver/HStore.java       |   30 +-
 .../regionserver/RegionCoprocessorHost.java     |   10 +-
 .../apache/hadoop/hbase/regionserver/Store.java |    4 +-
 .../hadoop/hbase/regionserver/StoreFile.java    | 1099 +-----------------
 .../hbase/regionserver/StoreFileInfo.java       |    6 +-
 .../hbase/regionserver/StoreFileReader.java     |  647 +++++++++++
 .../hbase/regionserver/StoreFileScanner.java    |    9 +-
 .../hbase/regionserver/StoreFileWriter.java     |  514 ++++++++
 .../hadoop/hbase/regionserver/StoreFlusher.java |    2 +-
 .../hadoop/hbase/regionserver/StoreUtils.java   |    2 +-
 .../regionserver/StripeMultiFileWriter.java     |   15 +-
 .../hbase/regionserver/StripeStoreFlusher.java  |    6 +-
 .../AbstractMultiOutputCompactor.java           |    4 +-
 .../compactions/CompactionRequest.java          |    4 +-
 .../regionserver/compactions/Compactor.java     |    8 +-
 .../compactions/DefaultCompactor.java           |   26 +-
 .../hbase/coprocessor/SimpleRegionObserver.java |   10 +-
 .../hadoop/hbase/io/hfile/TestCacheOnWrite.java |    4 +-
 .../hadoop/hbase/io/hfile/TestPrefetch.java     |    4 +-
 .../hfile/TestSeekBeforeWithInlineBlocks.java   |    6 +-
 .../apache/hadoop/hbase/mob/MobTestUtil.java    |    6 +-
 .../hadoop/hbase/mob/TestCachedMobFile.java     |   10 +-
 .../apache/hadoop/hbase/mob/TestMobFile.java    |    5 +-
 .../hadoop/hbase/mob/TestMobFileCache.java      |    4 +-
 .../TestPartitionedMobCompactor.java            |    6 +-
 .../regionserver/CreateRandomStoreFile.java     |    2 +-
 .../regionserver/DataBlockEncodingTool.java     |    2 +-
 .../EncodedSeekPerformanceTest.java             |    4 +-
 .../hbase/regionserver/MockStoreFile.java       |    5 +-
 .../regionserver/TestCacheOnWriteInSchema.java  |    4 +-
 .../hbase/regionserver/TestCompaction.java      |    2 +-
 .../regionserver/TestCompoundBloomFilter.java   |    4 +-
 .../hbase/regionserver/TestFSErrorsExposed.java |    6 +-
 .../hbase/regionserver/TestHMobStore.java       |    2 +-
 .../hadoop/hbase/regionserver/TestHRegion.java  |    4 +-
 .../regionserver/TestReversibleScanners.java    |   18 +-
 .../regionserver/TestScanWithBloomError.java    |    2 +-
 .../hadoop/hbase/regionserver/TestStore.java    |    6 +-
 .../hbase/regionserver/TestStoreFile.java       |   60 +-
 .../TestStoreFileScannerWithTagCompression.java |    6 +-
 .../regionserver/TestStripeStoreEngine.java     |    2 +-
 .../compactions/MockStoreFileGenerator.java     |    3 +-
 .../regionserver/compactions/TestCompactor.java |   12 +-
 .../compactions/TestStripeCompactionPolicy.java |    3 +-
 .../hadoop/hbase/spark/HBaseContext.scala       |   10 +-
 61 files changed, 1412 insertions(+), 1332 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java
index da7252b..67093f8 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java
@@ -55,7 +55,7 @@ import org.apache.hadoop.hbase.regionserver.RegionScanner;
 import org.apache.hadoop.hbase.regionserver.ScanType;
 import org.apache.hadoop.hbase.regionserver.Store;
 import org.apache.hadoop.hbase.regionserver.StoreFile;
-import org.apache.hadoop.hbase.regionserver.StoreFile.Reader;
+import org.apache.hadoop.hbase.regionserver.StoreFileReader;
 import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
 import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
 import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
@@ -485,16 +485,16 @@ public class BaseRegionObserver implements RegionObserver 
{
   }
 
   @Override
-  public Reader 
preStoreFileReaderOpen(ObserverContext<RegionCoprocessorEnvironment> ctx,
+  public StoreFileReader 
preStoreFileReaderOpen(ObserverContext<RegionCoprocessorEnvironment> ctx,
       FileSystem fs, Path p, FSDataInputStreamWrapper in, long size, 
CacheConfig cacheConf,
-      Reference r, Reader reader) throws IOException {
+      Reference r, StoreFileReader reader) throws IOException {
     return reader;
   }
 
   @Override
-  public Reader 
postStoreFileReaderOpen(ObserverContext<RegionCoprocessorEnvironment> ctx,
+  public StoreFileReader 
postStoreFileReaderOpen(ObserverContext<RegionCoprocessorEnvironment> ctx,
       FileSystem fs, Path p, FSDataInputStreamWrapper in, long size, 
CacheConfig cacheConf,
-      Reference r, Reader reader) throws IOException {
+      Reference r, StoreFileReader reader) throws IOException {
     return reader;
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java
index e133f73..db08b76 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java
@@ -55,6 +55,7 @@ import org.apache.hadoop.hbase.regionserver.RegionScanner;
 import org.apache.hadoop.hbase.regionserver.ScanType;
 import org.apache.hadoop.hbase.regionserver.Store;
 import org.apache.hadoop.hbase.regionserver.StoreFile;
+import org.apache.hadoop.hbase.regionserver.StoreFileReader;
 import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
 import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
 import org.apache.hadoop.hbase.wal.WALKey;
@@ -1210,9 +1211,9 @@ public interface RegionObserver extends Coprocessor {
    * default behavior, null otherwise
    * @throws IOException
    */
-  StoreFile.Reader preStoreFileReaderOpen(final 
ObserverContext<RegionCoprocessorEnvironment> ctx,
+  StoreFileReader preStoreFileReaderOpen(final 
ObserverContext<RegionCoprocessorEnvironment> ctx,
       final FileSystem fs, final Path p, final FSDataInputStreamWrapper in, 
long size,
-      final CacheConfig cacheConf, final Reference r, StoreFile.Reader reader) 
throws IOException;
+      final CacheConfig cacheConf, final Reference r, StoreFileReader reader) 
throws IOException;
 
   /**
    * Called after the creation of Reader for a store file.
@@ -1228,9 +1229,9 @@ public interface RegionObserver extends Coprocessor {
    * @return The reader to use
    * @throws IOException
    */
-  StoreFile.Reader postStoreFileReaderOpen(final 
ObserverContext<RegionCoprocessorEnvironment> ctx,
+  StoreFileReader postStoreFileReaderOpen(final 
ObserverContext<RegionCoprocessorEnvironment> ctx,
       final FileSystem fs, final Path p, final FSDataInputStreamWrapper in, 
long size,
-      final CacheConfig cacheConf, final Reference r, StoreFile.Reader reader) 
throws IOException;
+      final CacheConfig cacheConf, final Reference r, StoreFileReader reader) 
throws IOException;
 
   /**
    * Called after a new cell has been created during an increment operation, 
but before

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java
index 067d24c..a4a281e 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java
@@ -33,7 +33,7 @@ import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.io.hfile.CacheConfig;
 import org.apache.hadoop.hbase.io.hfile.HFileScanner;
-import org.apache.hadoop.hbase.regionserver.StoreFile;
+import org.apache.hadoop.hbase.regionserver.StoreFileReader;
 import org.apache.hadoop.hbase.util.Bytes;
 
 /**
@@ -50,7 +50,7 @@ import org.apache.hadoop.hbase.util.Bytes;
  * <p>This file is not splitable.  Calls to {@link #midkey()} return null.
  */
 @InterfaceAudience.Private
-public class HalfStoreFileReader extends StoreFile.Reader {
+public class HalfStoreFileReader extends StoreFileReader {
   private static final Log LOG = LogFactory.getLog(HalfStoreFileReader.class);
   final boolean top;
   // This is the key we split around.  Its the first possible entry on a row:

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java
index cd60562..d95aafd 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CompoundBloomFilterWriter.java
@@ -29,6 +29,7 @@ import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.hbase.classification.InterfaceAudience;
 import org.apache.hadoop.hbase.CellComparator;
+import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
 import org.apache.hadoop.hbase.util.BloomFilterChunk;
 import org.apache.hadoop.hbase.util.BloomFilterUtil;
 import org.apache.hadoop.hbase.util.BloomFilterWriter;
@@ -155,7 +156,7 @@ public class CompoundBloomFilterWriter extends 
CompoundBloomFilterBase
    * Adds a Bloom filter key. This key must be greater than the previous key,
    * as defined by the comparator this compound Bloom filter is configured
    * with. For efficiency, key monotonicity is not checked here. See
-   * {@link org.apache.hadoop.hbase.regionserver.StoreFile.Writer#append(
+   * {@link StoreFileWriter#append(
    * org.apache.hadoop.hbase.Cell)} for the details of deduplication.
    */
   @Override

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
index 059e030..606d336 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java
@@ -65,6 +65,7 @@ import org.apache.hadoop.hbase.io.hfile.HFileContextBuilder;
 import org.apache.hadoop.hbase.regionserver.BloomType;
 import org.apache.hadoop.hbase.regionserver.HStore;
 import org.apache.hadoop.hbase.regionserver.StoreFile;
+import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.io.NullWritable;
 import org.apache.hadoop.io.SequenceFile;
@@ -305,12 +306,12 @@ public class HFileOutputFormat2
                                     
         if (null == favoredNodes) {
           wl.writer =
-              new StoreFile.WriterBuilder(conf, new CacheConfig(tempConf), fs)
+              new StoreFileWriter.Builder(conf, new CacheConfig(tempConf), fs)
                   .withOutputDir(familydir).withBloomType(bloomType)
                   
.withComparator(CellComparator.COMPARATOR).withFileContext(hFileContext).build();
         } else {
           wl.writer =
-              new StoreFile.WriterBuilder(conf, new CacheConfig(tempConf), new 
HFileSystem(fs))
+              new StoreFileWriter.Builder(conf, new CacheConfig(tempConf), new 
HFileSystem(fs))
                   .withOutputDir(familydir).withBloomType(bloomType)
                   
.withComparator(CellComparator.COMPARATOR).withFileContext(hFileContext)
                   .withFavoredNodes(favoredNodes).build();
@@ -320,7 +321,7 @@ public class HFileOutputFormat2
         return wl;
       }
 
-      private void close(final StoreFile.Writer w) throws IOException {
+      private void close(final StoreFileWriter w) throws IOException {
         if (w != null) {
           w.appendFileInfo(StoreFile.BULKLOAD_TIME_KEY,
               Bytes.toBytes(System.currentTimeMillis()));
@@ -350,7 +351,7 @@ public class HFileOutputFormat2
    */
   static class WriterLength {
     long written = 0;
-    StoreFile.Writer writer = null;
+    StoreFileWriter writer = null;
   }
 
   /**

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java
index 5d75d56..86a84a4 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java
@@ -65,8 +65,8 @@ import org.apache.hadoop.hbase.io.hfile.HFileScanner;
 import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.regionserver.BloomType;
 import org.apache.hadoop.hbase.regionserver.HStore;
-import org.apache.hadoop.hbase.regionserver.StoreFile;
 import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
+import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
 import org.apache.hadoop.hbase.security.UserProvider;
 import org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint;
 import org.apache.hadoop.hbase.security.token.FsDelegationToken;
@@ -890,7 +890,7 @@ public class LoadIncrementalHFiles extends Configured 
implements Tool {
     FileSystem fs = inFile.getFileSystem(conf);
     CacheConfig cacheConf = new CacheConfig(conf);
     HalfStoreFileReader halfReader = null;
-    StoreFile.Writer halfWriter = null;
+    StoreFileWriter halfWriter = null;
     try {
       halfReader = new HalfStoreFileReader(fs, inFile, cacheConf, reference, 
conf);
       Map<byte[], byte[]> fileInfo = halfReader.loadFileInfo();
@@ -906,7 +906,7 @@ public class LoadIncrementalHFiles extends Configured 
implements Tool {
                                   
.withDataBlockEncoding(familyDescriptor.getDataBlockEncoding())
                                   .withIncludesTags(true)
                                   .build();
-      halfWriter = new StoreFile.WriterBuilder(conf, cacheConf,
+      halfWriter = new StoreFileWriter.Builder(conf, cacheConf,
           fs)
               .withFilePath(outFile)
               .withBloomType(bloomFilterType)

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
index fe640c6..711f31d 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
@@ -42,8 +42,8 @@ import 
org.apache.hadoop.hbase.regionserver.MobCompactionStoreScanner;
 import org.apache.hadoop.hbase.regionserver.ScanType;
 import org.apache.hadoop.hbase.regionserver.ScannerContext;
 import org.apache.hadoop.hbase.regionserver.Store;
-import org.apache.hadoop.hbase.regionserver.StoreFile.Writer;
 import org.apache.hadoop.hbase.regionserver.StoreFileScanner;
+import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
 import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
 import org.apache.hadoop.hbase.regionserver.compactions.DefaultCompactor;
 import org.apache.hadoop.hbase.regionserver.throttle.ThroughputController;
@@ -86,17 +86,17 @@ public class DefaultMobStoreCompactor extends 
DefaultCompactor {
     }
   };
 
-  private final CellSinkFactory<Writer> writerFactory = new 
CellSinkFactory<Writer>() {
-
-    @Override
-    public Writer createWriter(InternalScanner scanner,
-        org.apache.hadoop.hbase.regionserver.compactions.Compactor.FileDetails 
fd,
-        boolean shouldDropBehind) throws IOException {
-      // make this writer with tags always because of possible new cells with 
tags.
-      return store.createWriterInTmp(fd.maxKeyCount, compactionCompression, 
true, true, true,
-        shouldDropBehind);
-    }
-  };
+  private final CellSinkFactory<StoreFileWriter> writerFactory =
+      new CellSinkFactory<StoreFileWriter>() {
+        @Override
+        public StoreFileWriter createWriter(InternalScanner scanner,
+            
org.apache.hadoop.hbase.regionserver.compactions.Compactor.FileDetails fd,
+            boolean shouldDropBehind) throws IOException {
+          // make this writer with tags always because of possible new cells 
with tags.
+          return store.createWriterInTmp(fd.maxKeyCount, 
compactionCompression, true, true, true,
+            shouldDropBehind);
+        }
+      };
 
   public DefaultMobStoreCompactor(Configuration conf, Store store) {
     super(conf, store);
@@ -180,7 +180,7 @@ public class DefaultMobStoreCompactor extends 
DefaultCompactor {
     boolean hasMore;
     Path path = MobUtils.getMobFamilyPath(conf, store.getTableName(), 
store.getColumnFamilyName());
     byte[] fileName = null;
-    Writer mobFileWriter = null, delFileWriter = null;
+    StoreFileWriter mobFileWriter = null, delFileWriter = null;
     long mobCells = 0, deleteMarkersCount = 0;
     Tag tableNameTag = new ArrayBackedTag(TagType.MOB_TABLE_NAME_TAG_TYPE,
         store.getTableName().getName());

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java
index eb6c739..04a8782 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java
@@ -41,7 +41,7 @@ import org.apache.hadoop.hbase.regionserver.InternalScanner;
 import org.apache.hadoop.hbase.regionserver.MemStoreSnapshot;
 import org.apache.hadoop.hbase.regionserver.ScannerContext;
 import org.apache.hadoop.hbase.regionserver.Store;
-import org.apache.hadoop.hbase.regionserver.StoreFile;
+import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
 import org.apache.hadoop.hbase.regionserver.throttle.ThroughputController;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.util.StringUtils;
@@ -108,7 +108,7 @@ public class DefaultMobStoreFlusher extends 
DefaultStoreFlusher {
     if (scanner == null) {
       return result; // NULL scanner returned from coprocessor hooks means 
skip normal processing
     }
-    StoreFile.Writer writer;
+    StoreFileWriter writer;
     try {
       // TODO: We can fail in the below block before we complete adding this 
flush to
       // list of store files. Add cleanup of anything put on filesystem if we 
fail.
@@ -155,8 +155,8 @@ public class DefaultMobStoreFlusher extends 
DefaultStoreFlusher {
    * @throws IOException
    */
   protected void performMobFlush(MemStoreSnapshot snapshot, long cacheFlushId,
-      InternalScanner scanner, StoreFile.Writer writer, MonitoredTask status) 
throws IOException {
-    StoreFile.Writer mobFileWriter = null;
+      InternalScanner scanner, StoreFileWriter writer, MonitoredTask status) 
throws IOException {
+    StoreFileWriter mobFileWriter = null;
     int compactionKVMax = conf.getInt(HConstants.COMPACTION_KV_MAX,
         HConstants.COMPACTION_KV_MAX_DEFAULT);
     long mobCount = 0;

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java
index 52a19f5..ecd2415 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java
@@ -68,6 +68,7 @@ import 
org.apache.hadoop.hbase.mob.compactions.PartitionedMobCompactor;
 import org.apache.hadoop.hbase.regionserver.BloomType;
 import org.apache.hadoop.hbase.regionserver.HStore;
 import org.apache.hadoop.hbase.regionserver.StoreFile;
+import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
 import org.apache.hadoop.hbase.util.FSUtils;
@@ -462,7 +463,7 @@ public final class MobUtils {
    * @return The writer for the mob file.
    * @throws IOException
    */
-  public static StoreFile.Writer createWriter(Configuration conf, FileSystem 
fs,
+  public static StoreFileWriter createWriter(Configuration conf, FileSystem fs,
       HColumnDescriptor family, String date, Path basePath, long maxKeyCount,
       Compression.Algorithm compression, String startKey, CacheConfig 
cacheConfig,
       Encryption.Context cryptoContext)
@@ -485,7 +486,7 @@ public final class MobUtils {
    * @return The writer for the mob file.
    * @throws IOException
    */
-  public static StoreFile.Writer createRefFileWriter(Configuration conf, 
FileSystem fs,
+  public static StoreFileWriter createRefFileWriter(Configuration conf, 
FileSystem fs,
     HColumnDescriptor family, Path basePath, long maxKeyCount, CacheConfig 
cacheConfig,
     Encryption.Context cryptoContext)
     throws IOException {
@@ -497,7 +498,7 @@ public final class MobUtils {
       
.withEncryptionContext(cryptoContext).withCreateTime(EnvironmentEdgeManager.currentTime())
       .build();
     Path tempPath = new Path(basePath, 
UUID.randomUUID().toString().replaceAll("-", ""));
-    StoreFile.Writer w = new StoreFile.WriterBuilder(conf, cacheConfig, 
fs).withFilePath(tempPath)
+    StoreFileWriter w = new StoreFileWriter.Builder(conf, cacheConfig, 
fs).withFilePath(tempPath)
       
.withComparator(CellComparator.COMPARATOR).withBloomType(family.getBloomFilterType())
       .withMaxKeyCount(maxKeyCount).withFileContext(hFileContext).build();
     return w;
@@ -518,7 +519,7 @@ public final class MobUtils {
    * @return The writer for the mob file.
    * @throws IOException
    */
-  public static StoreFile.Writer createWriter(Configuration conf, FileSystem 
fs,
+  public static StoreFileWriter createWriter(Configuration conf, FileSystem fs,
       HColumnDescriptor family, String date, Path basePath, long maxKeyCount,
       Compression.Algorithm compression, byte[] startKey, CacheConfig 
cacheConfig,
       Encryption.Context cryptoContext)
@@ -544,7 +545,7 @@ public final class MobUtils {
    * @return The writer for the del file.
    * @throws IOException
    */
-  public static StoreFile.Writer createDelFileWriter(Configuration conf, 
FileSystem fs,
+  public static StoreFileWriter createDelFileWriter(Configuration conf, 
FileSystem fs,
       HColumnDescriptor family, String date, Path basePath, long maxKeyCount,
       Compression.Algorithm compression, byte[] startKey, CacheConfig 
cacheConfig,
       Encryption.Context cryptoContext)
@@ -570,7 +571,7 @@ public final class MobUtils {
    * @return The writer for the mob file.
    * @throws IOException
    */
-  private static StoreFile.Writer createWriter(Configuration conf, FileSystem 
fs,
+  private static StoreFileWriter createWriter(Configuration conf, FileSystem 
fs,
     HColumnDescriptor family, MobFileName mobFileName, Path basePath, long 
maxKeyCount,
     Compression.Algorithm compression, CacheConfig cacheConfig, 
Encryption.Context cryptoContext)
     throws IOException {
@@ -583,7 +584,7 @@ public final class MobUtils {
       .withEncryptionContext(cryptoContext)
       .withCreateTime(EnvironmentEdgeManager.currentTime()).build();
 
-    StoreFile.Writer w = new StoreFile.WriterBuilder(conf, cacheConfig, fs)
+    StoreFileWriter w = new StoreFileWriter.Builder(conf, cacheConfig, fs)
       .withFilePath(new Path(basePath, mobFileName.getFileName()))
       .withComparator(CellComparator.COMPARATOR).withBloomType(BloomType.NONE)
       .withMaxKeyCount(maxKeyCount).withFileContext(hFileContext).build();

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java
index 6c6f115..cfe76ae 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactor.java
@@ -69,9 +69,9 @@ import org.apache.hadoop.hbase.regionserver.ScanInfo;
 import org.apache.hadoop.hbase.regionserver.ScanType;
 import org.apache.hadoop.hbase.regionserver.ScannerContext;
 import org.apache.hadoop.hbase.regionserver.StoreFile;
-import org.apache.hadoop.hbase.regionserver.StoreFile.Writer;
 import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
 import org.apache.hadoop.hbase.regionserver.StoreFileScanner;
+import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
 import org.apache.hadoop.hbase.regionserver.StoreScanner;
 import org.apache.hadoop.hbase.security.EncryptionUtil;
 import org.apache.hadoop.hbase.util.Bytes;
@@ -385,8 +385,8 @@ public class PartitionedMobCompactor extends MobCompactor {
     // Pair(maxSeqId, cellsCount)
     Pair<Long, Long> fileInfo = getFileInfo(mobFilesToCompact);
     // open writers for the mob files and new ref store files.
-    Writer writer = null;
-    Writer refFileWriter = null;
+    StoreFileWriter writer = null;
+    StoreFileWriter refFileWriter = null;
     Path filePath = null;
     Path refFilePath = null;
     long mobCells = 0;
@@ -499,7 +499,7 @@ public class PartitionedMobCompactor extends MobCompactor {
     List<StoreFile> delFiles) throws IOException {
     // create a scanner for the del files.
     StoreScanner scanner = createScanner(delFiles, 
ScanType.COMPACT_RETAIN_DELETES);
-    Writer writer = null;
+    StoreFileWriter writer = null;
     Path filePath = null;
     try {
       writer = MobUtils.createDelFileWriter(conf, fs, column,
@@ -589,7 +589,7 @@ public class PartitionedMobCompactor extends MobCompactor {
    * @param mobCellsCount The number of mob cells.
    * @throws IOException
    */
-  private void closeMobFileWriter(Writer writer, long maxSeqId, long 
mobCellsCount)
+  private void closeMobFileWriter(StoreFileWriter writer, long maxSeqId, long 
mobCellsCount)
     throws IOException {
     if (writer != null) {
       writer.appendMetadata(maxSeqId, false, mobCellsCount);
@@ -608,7 +608,7 @@ public class PartitionedMobCompactor extends MobCompactor {
    * @param bulkloadTime The timestamp at which the bulk load file is created.
    * @throws IOException
    */
-  private void closeRefFileWriter(Writer writer, long maxSeqId, long 
bulkloadTime)
+  private void closeRefFileWriter(StoreFileWriter writer, long maxSeqId, long 
bulkloadTime)
     throws IOException {
     if (writer != null) {
       writer.appendMetadata(maxSeqId, false);

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java
index 5955cc2..7997b49 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/mapreduce/MemStoreWrapper.java
@@ -45,7 +45,7 @@ import 
org.apache.hadoop.hbase.mob.mapreduce.SweepJob.SweepCounter;
 import org.apache.hadoop.hbase.regionserver.KeyValueScanner;
 import org.apache.hadoop.hbase.regionserver.MemStore;
 import org.apache.hadoop.hbase.regionserver.MemStoreSnapshot;
-import org.apache.hadoop.hbase.regionserver.StoreFile;
+import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
 import org.apache.hadoop.hbase.security.EncryptionUtil;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.mapreduce.Reducer.Context;
@@ -132,7 +132,7 @@ public class MemStoreWrapper {
     }
     // generate the files into a temp directory.
     String tempPathString = 
context.getConfiguration().get(SweepJob.WORKING_FILES_DIR_KEY);
-    StoreFile.Writer mobFileWriter = MobUtils.createWriter(conf, fs, hcd, 
partitionId.getDate(),
+    StoreFileWriter mobFileWriter = MobUtils.createWriter(conf, fs, hcd, 
partitionId.getDate(),
       new Path(tempPathString), snapshot.getCellsCount(), 
hcd.getCompactionCompression(),
       partitionId.getStartKey(), cacheConfig, cryptoContext);
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java
index 4987c59..a4e0285 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java
@@ -26,7 +26,6 @@ import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.regionserver.StoreFile.Writer;
 import org.apache.hadoop.hbase.regionserver.compactions.Compactor.CellSink;
 
 /**
@@ -44,7 +43,7 @@ public abstract class AbstractMultiFileWriter implements 
CellSink {
   protected StoreScanner sourceScanner;
 
   public interface WriterFactory {
-    public StoreFile.Writer createWriter() throws IOException;
+    public StoreFileWriter createWriter() throws IOException;
   }
 
   /**
@@ -66,13 +65,13 @@ public abstract class AbstractMultiFileWriter implements 
CellSink {
    */
   public List<Path> commitWriters(long maxSeqId, boolean majorCompaction) 
throws IOException {
     preCommitWriters();
-    Collection<StoreFile.Writer> writers = this.writers();
+    Collection<StoreFileWriter> writers = this.writers();
     if (LOG.isDebugEnabled()) {
       LOG.debug("Commit " + writers.size() + " writers, maxSeqId=" + maxSeqId
           + ", majorCompaction=" + majorCompaction);
     }
     List<Path> paths = new ArrayList<Path>();
-    for (Writer writer : writers) {
+    for (StoreFileWriter writer : writers) {
       if (writer == null) {
         continue;
       }
@@ -89,7 +88,7 @@ public abstract class AbstractMultiFileWriter implements 
CellSink {
    */
   public List<Path> abortWriters() {
     List<Path> paths = new ArrayList<Path>();
-    for (StoreFile.Writer writer : writers()) {
+    for (StoreFileWriter writer : writers()) {
       try {
         if (writer != null) {
           paths.add(writer.getPath());
@@ -102,7 +101,7 @@ public abstract class AbstractMultiFileWriter implements 
CellSink {
     return paths;
   }
 
-  protected abstract Collection<StoreFile.Writer> writers();
+  protected abstract Collection<StoreFileWriter> writers();
 
   /**
    * Subclasses override this method to be called at the end of a successful 
sequence of append; all
@@ -115,6 +114,6 @@ public abstract class AbstractMultiFileWriter implements 
CellSink {
    * Subclasses override this method to be called before we close the give 
writer. Usually you can
    * append extra metadata to the writer.
    */
-  protected void preCloseWriter(StoreFile.Writer writer) throws IOException {
+  protected void preCloseWriter(StoreFileWriter writer) throws IOException {
   }
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java
index f0bd444..2cea92f 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DateTieredMultiFileWriter.java
@@ -26,7 +26,6 @@ import java.util.TreeMap;
 
 import org.apache.hadoop.hbase.Cell;
 import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.regionserver.StoreFile.Writer;
 
 /**
  * class for cell sink that separates the provided cells into multiple files 
for date tiered
@@ -35,8 +34,8 @@ import org.apache.hadoop.hbase.regionserver.StoreFile.Writer;
 @InterfaceAudience.Private
 public class DateTieredMultiFileWriter extends AbstractMultiFileWriter {
 
-  private final NavigableMap<Long, StoreFile.Writer> lowerBoundary2Writer
-    = new TreeMap<Long, StoreFile.Writer>();
+  private final NavigableMap<Long, StoreFileWriter> lowerBoundary2Writer
+    = new TreeMap<Long, StoreFileWriter>();
 
   private final boolean needEmptyFile;
 
@@ -53,8 +52,8 @@ public class DateTieredMultiFileWriter extends 
AbstractMultiFileWriter {
 
   @Override
   public void append(Cell cell) throws IOException {
-    Map.Entry<Long, StoreFile.Writer> entry = 
lowerBoundary2Writer.floorEntry(cell.getTimestamp());
-    StoreFile.Writer writer = entry.getValue();
+    Map.Entry<Long, StoreFileWriter> entry = 
lowerBoundary2Writer.floorEntry(cell.getTimestamp());
+    StoreFileWriter writer = entry.getValue();
     if (writer == null) {
       writer = writerFactory.createWriter();
       lowerBoundary2Writer.put(entry.getKey(), writer);
@@ -63,7 +62,7 @@ public class DateTieredMultiFileWriter extends 
AbstractMultiFileWriter {
   }
 
   @Override
-  protected Collection<Writer> writers() {
+  protected Collection<StoreFileWriter> writers() {
     return lowerBoundary2Writer.values();
   }
 
@@ -72,7 +71,7 @@ public class DateTieredMultiFileWriter extends 
AbstractMultiFileWriter {
     if (!needEmptyFile) {
       return;
     }
-    for (StoreFile.Writer writer : lowerBoundary2Writer.values()) {
+    for (StoreFileWriter writer : lowerBoundary2Writer.values()) {
       if (writer != null) {
         return;
       }

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java
index 711b987..bf8794a 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/DefaultStoreFlusher.java
@@ -57,7 +57,7 @@ public class DefaultStoreFlusher extends StoreFlusher {
       return result; // NULL scanner returned from coprocessor hooks means 
skip normal processing
     }
 
-    StoreFile.Writer writer;
+    StoreFileWriter writer;
     try {
       // TODO:  We can fail in the below block before we complete adding this 
flush to
       //        list of store files.  Add cleanup of anything put on 
filesystem if we fail.

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java
index 7b44338..54b942d 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HMobStore.java
@@ -188,7 +188,7 @@ public class HMobStore extends HStore {
    * @return The writer for the mob file.
    * @throws IOException
    */
-  public StoreFile.Writer createWriterInTmp(Date date, long maxKeyCount,
+  public StoreFileWriter createWriterInTmp(Date date, long maxKeyCount,
       Compression.Algorithm compression, byte[] startKey) throws IOException {
     if (startKey == null) {
       startKey = HConstants.EMPTY_START_ROW;
@@ -208,7 +208,7 @@ public class HMobStore extends HStore {
    * @return The writer for the del file.
    * @throws IOException
    */
-  public StoreFile.Writer createDelFileWriterInTmp(Date date, long maxKeyCount,
+  public StoreFileWriter createDelFileWriterInTmp(Date date, long maxKeyCount,
       Compression.Algorithm compression, byte[] startKey) throws IOException {
     if (startKey == null) {
       startKey = HConstants.EMPTY_START_ROW;
@@ -230,7 +230,7 @@ public class HMobStore extends HStore {
    * @return The writer for the mob file.
    * @throws IOException
    */
-  public StoreFile.Writer createWriterInTmp(String date, Path basePath, long 
maxKeyCount,
+  public StoreFileWriter createWriterInTmp(String date, Path basePath, long 
maxKeyCount,
       Compression.Algorithm compression, byte[] startKey) throws IOException {
     MobFileName mobFileName = MobFileName.create(startKey, date, 
UUID.randomUUID()
         .toString().replaceAll("-", ""));
@@ -246,7 +246,7 @@ public class HMobStore extends HStore {
    * @return The writer for the mob file.
    * @throws IOException
    */
-  public StoreFile.Writer createWriterInTmp(MobFileName mobFileName, Path 
basePath,
+  public StoreFileWriter createWriterInTmp(MobFileName mobFileName, Path 
basePath,
       long maxKeyCount, Compression.Algorithm compression) throws IOException {
     final CacheConfig writerCacheConf = mobCacheConfig;
     HFileContext hFileContext = new 
HFileContextBuilder().withCompression(compression)
@@ -259,7 +259,7 @@ public class HMobStore extends HStore {
         .withEncryptionContext(cryptoContext)
         .withCreateTime(EnvironmentEdgeManager.currentTime()).build();
 
-    StoreFile.Writer w = new StoreFile.WriterBuilder(conf, writerCacheConf, 
region.getFilesystem())
+    StoreFileWriter w = new StoreFileWriter.Builder(conf, writerCacheConf, 
region.getFilesystem())
         .withFilePath(new Path(basePath, mobFileName.getFileName()))
         
.withComparator(CellComparator.COMPARATOR).withBloomType(BloomType.NONE)
         .withMaxKeyCount(maxKeyCount).withFileContext(hFileContext).build();

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
index 5524de0..72238cc 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
@@ -1691,7 +1691,7 @@ public class HRegion implements HeapSize, 
PropagatingConfigurationObserver, Regi
       Collection<StoreFile> storeFiles = store.getStorefiles();
       if (storeFiles == null) continue;
       for (StoreFile file : storeFiles) {
-        StoreFile.Reader sfReader = file.getReader();
+        StoreFileReader sfReader = file.getReader();
         if (sfReader == null) continue;
         HFile.Reader reader = sfReader.getHFileReader();
         if (reader == null) continue;

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java
index f7ad4ae..7468be0 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java
@@ -611,7 +611,7 @@ public class HStore implements Store {
     info.setRegionCoprocessorHost(this.region.getCoprocessorHost());
     StoreFile storeFile = new StoreFile(this.getFileSystem(), info, this.conf, 
this.cacheConf,
       this.family.getBloomFilterType());
-    StoreFile.Reader r = storeFile.createReader();
+    StoreFileReader r = storeFile.createReader();
     r.setReplicaStoreFile(isPrimaryReplicaStore());
     return storeFile;
   }
@@ -749,7 +749,7 @@ public class HStore implements Store {
   }
 
   private void bulkLoadHFile(StoreFile sf) throws IOException {
-    StoreFile.Reader r = sf.getReader();
+    StoreFileReader r = sf.getReader();
     this.storeSize += r.length();
     this.totalUncompressedBytes += r.getTotalUncompressedBytes();
 
@@ -917,7 +917,7 @@ public class HStore implements Store {
     status.setStatus("Flushing " + this + ": reopening flushed file");
     StoreFile sf = createStoreFileAndReader(dstPath);
 
-    StoreFile.Reader r = sf.getReader();
+    StoreFileReader r = sf.getReader();
     this.storeSize += r.length();
     this.totalUncompressedBytes += r.getTotalUncompressedBytes();
 
@@ -930,7 +930,7 @@ public class HStore implements Store {
   }
 
   @Override
-  public StoreFile.Writer createWriterInTmp(long maxKeyCount, 
Compression.Algorithm compression,
+  public StoreFileWriter createWriterInTmp(long maxKeyCount, 
Compression.Algorithm compression,
                                             boolean isCompaction, boolean 
includeMVCCReadpoint,
                                             boolean includesTag)
       throws IOException {
@@ -947,7 +947,7 @@ public class HStore implements Store {
    * @return Writer for a new StoreFile in the tmp dir.
    */
   @Override
-  public StoreFile.Writer createWriterInTmp(long maxKeyCount, 
Compression.Algorithm compression,
+  public StoreFileWriter createWriterInTmp(long maxKeyCount, 
Compression.Algorithm compression,
       boolean isCompaction, boolean includeMVCCReadpoint, boolean includesTag,
       boolean shouldDropBehind)
   throws IOException {
@@ -966,7 +966,7 @@ public class HStore implements Store {
     }
     HFileContext hFileContext = createFileContext(compression, 
includeMVCCReadpoint, includesTag,
       cryptoContext);
-    StoreFile.Writer w = new StoreFile.WriterBuilder(conf, writerCacheConf,
+    StoreFileWriter w = new StoreFileWriter.Builder(conf, writerCacheConf,
         this.getFileSystem())
             .withFilePath(fs.createTempName())
             .withComparator(comparator)
@@ -1749,7 +1749,7 @@ public class HStore implements Store {
     this.storeSize = 0L;
     this.totalUncompressedBytes = 0L;
     for (StoreFile hsf : 
this.storeEngine.getStoreFileManager().getStorefiles()) {
-      StoreFile.Reader r = hsf.getReader();
+      StoreFileReader r = hsf.getReader();
       if (r == null) {
         LOG.warn("StoreFile " + hsf + " has a null Reader");
         continue;
@@ -1900,7 +1900,7 @@ public class HStore implements Store {
   public long getMaxStoreFileAge() {
     long earliestTS = Long.MAX_VALUE;
     for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) {
-      StoreFile.Reader r = s.getReader();
+      StoreFileReader r = s.getReader();
       if (r == null) {
         LOG.warn("StoreFile " + s + " has a null Reader");
         continue;
@@ -1919,7 +1919,7 @@ public class HStore implements Store {
   public long getMinStoreFileAge() {
     long latestTS = 0;
     for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) {
-      StoreFile.Reader r = s.getReader();
+      StoreFileReader r = s.getReader();
       if (r == null) {
         LOG.warn("StoreFile " + s + " has a null Reader");
         continue;
@@ -1938,7 +1938,7 @@ public class HStore implements Store {
   public long getAvgStoreFileAge() {
     long sum = 0, count = 0;
     for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) {
-      StoreFile.Reader r = s.getReader();
+      StoreFileReader r = s.getReader();
       if (r == null) {
         LOG.warn("StoreFile " + s + " has a null Reader");
         continue;
@@ -1988,7 +1988,7 @@ public class HStore implements Store {
   public long getStorefilesSize() {
     long size = 0;
     for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) {
-      StoreFile.Reader r = s.getReader();
+      StoreFileReader r = s.getReader();
       if (r == null) {
         LOG.warn("StoreFile " + s + " has a null Reader");
         continue;
@@ -2002,7 +2002,7 @@ public class HStore implements Store {
   public long getStorefilesIndexSize() {
     long size = 0;
     for (StoreFile s: this.storeEngine.getStoreFileManager().getStorefiles()) {
-      StoreFile.Reader r = s.getReader();
+      StoreFileReader r = s.getReader();
       if (r == null) {
         LOG.warn("StoreFile " + s + " has a null Reader");
         continue;
@@ -2016,7 +2016,7 @@ public class HStore implements Store {
   public long getTotalStaticIndexSize() {
     long size = 0;
     for (StoreFile s : this.storeEngine.getStoreFileManager().getStorefiles()) 
{
-      StoreFile.Reader r = s.getReader();
+      StoreFileReader r = s.getReader();
       if (r == null) {
         continue;
       }
@@ -2029,7 +2029,7 @@ public class HStore implements Store {
   public long getTotalStaticBloomSize() {
     long size = 0;
     for (StoreFile s : this.storeEngine.getStoreFileManager().getStorefiles()) 
{
-      StoreFile.Reader r = s.getReader();
+      StoreFileReader r = s.getReader();
       if (r == null) {
         continue;
       }
@@ -2430,7 +2430,7 @@ public class HStore implements Store {
     for (final StoreFile file : compactedfiles) {
       synchronized (file) {
         try {
-          StoreFile.Reader r = file.getReader();
+          StoreFileReader r = file.getReader();
           if (r == null) {
             if (LOG.isDebugEnabled()) {
               LOG.debug("The file " + file + " was closed but still not 
archived.");

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
index 48e67f7..6614b3e 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
@@ -1541,11 +1541,11 @@ public class RegionCoprocessorHost
    * default behavior, null otherwise
    * @throws IOException
    */
-  public StoreFile.Reader preStoreFileReaderOpen(final FileSystem fs, final 
Path p,
+  public StoreFileReader preStoreFileReaderOpen(final FileSystem fs, final 
Path p,
       final FSDataInputStreamWrapper in, final long size, final CacheConfig 
cacheConf,
       final Reference r) throws IOException {
     return execOperationWithResult(null,
-        coprocessors.isEmpty() ? null : new 
RegionOperationWithResult<StoreFile.Reader>() {
+        coprocessors.isEmpty() ? null : new 
RegionOperationWithResult<StoreFileReader>() {
       @Override
       public void call(RegionObserver oserver, 
ObserverContext<RegionCoprocessorEnvironment> ctx)
           throws IOException {
@@ -1565,11 +1565,11 @@ public class RegionCoprocessorHost
    * @return The reader to use
    * @throws IOException
    */
-  public StoreFile.Reader postStoreFileReaderOpen(final FileSystem fs, final 
Path p,
+  public StoreFileReader postStoreFileReaderOpen(final FileSystem fs, final 
Path p,
       final FSDataInputStreamWrapper in, final long size, final CacheConfig 
cacheConf,
-      final Reference r, final StoreFile.Reader reader) throws IOException {
+      final Reference r, final StoreFileReader reader) throws IOException {
     return execOperationWithResult(reader,
-        coprocessors.isEmpty() ? null : new 
RegionOperationWithResult<StoreFile.Reader>() {
+        coprocessors.isEmpty() ? null : new 
RegionOperationWithResult<StoreFileReader>() {
       @Override
       public void call(RegionObserver oserver, 
ObserverContext<RegionCoprocessorEnvironment> ctx)
           throws IOException {

http://git-wip-us.apache.org/repos/asf/hbase/blob/7efb9ede/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
----------------------------------------------------------------------
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
index 4cca4a1..b77a33b 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
@@ -164,7 +164,7 @@ public interface Store extends HeapSize, 
StoreConfigInformation, PropagatingConf
    * @param includeMVCCReadpoint whether we should out the MVCC readpoint
    * @return Writer for a new StoreFile in the tmp dir.
    */
-  StoreFile.Writer createWriterInTmp(
+  StoreFileWriter createWriterInTmp(
       long maxKeyCount,
       Compression.Algorithm compression,
       boolean isCompaction,
@@ -180,7 +180,7 @@ public interface Store extends HeapSize, 
StoreConfigInformation, PropagatingConf
    * @param shouldDropBehind should the writer drop caches behind writes
    * @return Writer for a new StoreFile in the tmp dir.
    */
-  StoreFile.Writer createWriterInTmp(
+  StoreFileWriter createWriterInTmp(
     long maxKeyCount,
     Compression.Algorithm compression,
     boolean isCompaction,

Reply via email to