HBASE-21263 Mention compression algorithm along with other storefile details
Signed-off-by: Andrew Purtell <apurt...@apache.org> Amending-Author: Andrew Purtell <apurt...@apache.org> Conflicts: hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/b50019d0 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/b50019d0 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/b50019d0 Branch: refs/heads/branch-1.2 Commit: b50019d0b3c8c906b5e2642fc4c330ff11a1f71d Parents: 56d38a0 Author: subrat.mishra <subrat.mis...@yahoo.com> Authored: Mon Oct 15 18:33:48 2018 +0530 Committer: Andrew Purtell <apurt...@apache.org> Committed: Tue Oct 16 13:41:02 2018 -0700 ---------------------------------------------------------------------- .../java/org/apache/hadoop/hbase/regionserver/HStore.java | 6 ++++++ .../hadoop/hbase/regionserver/compactions/Compactor.java | 1 + .../java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java | 9 +++++---- .../org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java | 7 ++++--- .../apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java | 2 +- .../hadoop/hbase/regionserver/CreateRandomStoreFile.java | 2 +- 6 files changed, 18 insertions(+), 9 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hbase/blob/b50019d0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java index 5de73b3..fb5790d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HStore.java @@ -331,6 +331,12 @@ public class HStore implements Store { cryptoContext.setCipher(cipher); cryptoContext.setKey(key); } + + LOG.info("Store=" + getColumnFamilyName() + + ", memstore type=" + this.memstore.getClass().getSimpleName() + + ", verifyBulkLoads=" + verifyBulkLoads + + ", encoding=" + family.getDataBlockEncoding() + + ", compression=" + family.getCompressionType()); } /** http://git-wip-us.apache.org/repos/asf/hbase/blob/b50019d0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java index f8fe2f0..dc79c25 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/Compactor.java @@ -174,6 +174,7 @@ public abstract class Compactor { ", bloomtype=" + r.getBloomFilterType().toString() + ", size=" + TraditionalBinaryPrefix.long2String(r.length(), "", 1) + ", encoding=" + r.getHFileReader().getDataBlockEncoding() + + ", compression=" + compactionCompression + ", seqNum=" + seqNum + (allFiles ? ", earliestPutTs=" + earliestPutTs: "")); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b50019d0/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java index 93477df..e404290 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestChecksum.java @@ -293,10 +293,11 @@ public class TestChecksum { dataSize + HConstants.HFILEBLOCK_HEADER_SIZE, bytesPerChecksum); LOG.info("testChecksumChunks: pread=" + pread + - ", bytesPerChecksum=" + bytesPerChecksum + - ", fileSize=" + totalSize + - ", dataSize=" + dataSize + - ", expectedChunks=" + expectedChunks); + ", bytesPerChecksum=" + bytesPerChecksum + + ", fileSize=" + totalSize + + ", dataSize=" + dataSize + + ", expectedChunks=" + expectedChunks + + ", compression=" + algo); // Verify hbase checksums. assertEquals(true, hfs.useHBaseChecksum()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b50019d0/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java index 9810448..1e6af02 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlock.java @@ -366,6 +366,8 @@ public class TestHFileBlock { for (Compression.Algorithm algo : COMPRESSION_ALGORITHMS) { for (boolean pread : new boolean[] { false, true }) { for (DataBlockEncoding encoding : DataBlockEncoding.values()) { + LOG.info("testDataBlockEncoding: Compression algorithm=" + algo + ", pread=" + pread + + ", dataBlockEncoder=" + encoding); Path path = new Path(TEST_UTIL.getDataTestDir(), "blocks_v2_" + algo + "_" + encoding.toString()); FSDataOutputStream os = fs.create(path); @@ -522,9 +524,8 @@ public class TestHFileBlock { for (boolean pread : BOOLEAN_VALUES) { for (boolean cacheOnWrite : BOOLEAN_VALUES) { Random rand = defaultRandom(); - LOG.info("testPreviousOffset:Compression algorithm: " + algo + - ", pread=" + pread + - ", cacheOnWrite=" + cacheOnWrite); + LOG.info("testPreviousOffset: Compression algorithm=" + algo + ", pread=" + pread + + ", cacheOnWrite=" + cacheOnWrite); Path path = new Path(TEST_UTIL.getDataTestDir(), "prev_offset"); List<Long> expectedOffsets = new ArrayList<Long>(); List<Long> expectedPrevOffsets = new ArrayList<Long>(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b50019d0/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java index 8c9353c..35008ec 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java @@ -189,7 +189,7 @@ public class TestHFileBlockIndex { private void readIndex(boolean useTags) throws IOException { long fileSize = fs.getFileStatus(path).getLen(); - LOG.info("Size of " + path + ": " + fileSize); + LOG.info("Size of " + path + ": " + fileSize + ", compression=" + compr); FSDataInputStream istream = fs.open(path); HFileContext meta = new HFileContextBuilder() http://git-wip-us.apache.org/repos/asf/hbase/blob/b50019d0/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/CreateRandomStoreFile.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/CreateRandomStoreFile.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/CreateRandomStoreFile.java index c3f7b82..bcba1f9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/CreateRandomStoreFile.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/CreateRandomStoreFile.java @@ -209,7 +209,7 @@ public class CreateRandomStoreFile { Path storeFilePath = sfw.getPath(); long fileSize = fs.getFileStatus(storeFilePath).getLen(); - LOG.info("Created " + storeFilePath + ", " + fileSize + " bytes"); + LOG.info("Created " + storeFilePath + ", " + fileSize + " bytes, compression=" + compr); return true; }