hbase git commit: HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result
Repository: hbase Updated Branches: refs/heads/branch-1.4 a9db7cf6f -> c8f2d61f0 HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c8f2d61f Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c8f2d61f Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c8f2d61f Branch: refs/heads/branch-1.4 Commit: c8f2d61f0a185c3998051860f1c87a8221b639a3 Parents: a9db7cf Author: Guanghao Zhang Authored: Thu Dec 20 11:03:54 2018 +0800 Committer: Guanghao Zhang Committed: Fri Dec 21 14:03:57 2018 +0800 -- .../hadoop/hbase/protobuf/ProtobufUtil.java | 4 +-- .../hbase/client/TestFromClientSide3.java | 4 +-- .../client/TestScannersFromClientSide.java | 38 .../hadoop/hbase/protobuf/TestProtobufUtil.java | 1 + 4 files changed, 42 insertions(+), 5 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/c8f2d61f/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index e630b98..4b516d4 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -1086,9 +1086,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/c8f2d61f/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java index 560f552..4cfb853 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java @@ -839,7 +839,7 @@ public class TestFromClientSide3 { } Scan scan = new Scan(); -scan.withStartRow(ROW).withStopRow(ROW).addFamily(FAMILY).setBatch(3) +scan.withStartRow(ROW).withStopRow(ROW, true).addFamily(FAMILY).setBatch(3) .setMaxResultSize(4 * 1024 * 1024); Result result; try (ResultScanner scanner = table.getScanner(scan)) { @@ -861,7 +861,7 @@ public class TestFromClientSide3 { } scan = new Scan(); -scan.withStartRow(ROW).withStopRow(ROW).addFamily(FAMILY).setBatch(2) +scan.withStartRow(ROW).withStopRow(ROW, true).addFamily(FAMILY).setBatch(2) .setMaxResultSize(4 * 1024 * 1024); try (ResultScanner scanner = table.getScanner(scan)) { List list = new ArrayList<>(); http://git-wip-us.apache.org/repos/asf/hbase/blob/c8f2d61f/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java index 17c8b92..4bac32d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java @@ -18,6 +18,7 @@ package org.apache.hadoop.hbase.client; import static org.junit.Assert.assertArrayEquals; import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertNotNull; import static org.junit.Assert.assertNull; import static org.junit.Assert.assertTrue; @@ -699,4 +700,41 @@ public class TestScannersFromClientSide { } } } + + @Test + public void testScanWithSameStartRowStopRow() throws IOException { +TableName tableName = TableName.valueOf("testScanWithSameStartRowStopRow"); +try (Table table = TEST_UTIL.createTable(tableName, FAMILY)) { + table.put(new Put(ROW).addColumn(FAMILY, QUALIFIER, VALUE)); + + Scan scan = new Scan().withStartRow(ROW).withStopRow(ROW); + try (ResultScanner scanner = table.getScanner(scan)) { +assertNull(scanner.next()); + } + + scan
hbase git commit: HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result
Repository: hbase Updated Branches: refs/heads/branch-1 bbf17e614 -> 98640e4b2 HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/98640e4b Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/98640e4b Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/98640e4b Branch: refs/heads/branch-1 Commit: 98640e4b2c895bcb9290c8f9e067e30ebe7cd257 Parents: bbf17e6 Author: Guanghao Zhang Authored: Thu Dec 20 11:03:54 2018 +0800 Committer: Guanghao Zhang Committed: Fri Dec 21 14:03:12 2018 +0800 -- .../hadoop/hbase/protobuf/ProtobufUtil.java | 4 +-- .../hbase/client/TestFromClientSide3.java | 4 +-- .../client/TestScannersFromClientSide.java | 38 .../hadoop/hbase/protobuf/TestProtobufUtil.java | 1 + 4 files changed, 42 insertions(+), 5 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/98640e4b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index e630b98..4b516d4 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -1086,9 +1086,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/98640e4b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java index 560f552..4cfb853 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java @@ -839,7 +839,7 @@ public class TestFromClientSide3 { } Scan scan = new Scan(); -scan.withStartRow(ROW).withStopRow(ROW).addFamily(FAMILY).setBatch(3) +scan.withStartRow(ROW).withStopRow(ROW, true).addFamily(FAMILY).setBatch(3) .setMaxResultSize(4 * 1024 * 1024); Result result; try (ResultScanner scanner = table.getScanner(scan)) { @@ -861,7 +861,7 @@ public class TestFromClientSide3 { } scan = new Scan(); -scan.withStartRow(ROW).withStopRow(ROW).addFamily(FAMILY).setBatch(2) +scan.withStartRow(ROW).withStopRow(ROW, true).addFamily(FAMILY).setBatch(2) .setMaxResultSize(4 * 1024 * 1024); try (ResultScanner scanner = table.getScanner(scan)) { List list = new ArrayList<>(); http://git-wip-us.apache.org/repos/asf/hbase/blob/98640e4b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java index 17c8b92..4bac32d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestScannersFromClientSide.java @@ -18,6 +18,7 @@ package org.apache.hadoop.hbase.client; import static org.junit.Assert.assertArrayEquals; import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertNotNull; import static org.junit.Assert.assertNull; import static org.junit.Assert.assertTrue; @@ -699,4 +700,41 @@ public class TestScannersFromClientSide { } } } + + @Test + public void testScanWithSameStartRowStopRow() throws IOException { +TableName tableName = TableName.valueOf("testScanWithSameStartRowStopRow"); +try (Table table = TEST_UTIL.createTable(tableName, FAMILY)) { + table.put(new Put(ROW).addColumn(FAMILY, QUALIFIER, VALUE)); + + Scan scan = new Scan().withStartRow(ROW).withStopRow(ROW); + try (ResultScanner scanner = table.getScanner(scan)) { +assertNull(scanner.next()); + } + + scan =
hbase git commit: HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result
Repository: hbase Updated Branches: refs/heads/branch-2.0 7e51e781f -> 9889ce9d3 HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/9889ce9d Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/9889ce9d Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/9889ce9d Branch: refs/heads/branch-2.0 Commit: 9889ce9d3446785253352733a00d6a68af01c21b Parents: 7e51e78 Author: Guanghao Zhang Authored: Thu Dec 20 11:03:54 2018 +0800 Committer: Guanghao Zhang Committed: Fri Dec 21 09:58:55 2018 +0800 -- .../hadoop/hbase/protobuf/ProtobufUtil.java | 4 +-- .../hbase/shaded/protobuf/ProtobufUtil.java | 4 +-- .../hbase/shaded/protobuf/TestProtobufUtil.java | 1 + .../hbase/client/TestFromClientSide3.java | 4 +-- .../client/TestScannersFromClientSide.java | 38 .../hadoop/hbase/protobuf/TestProtobufUtil.java | 1 + 6 files changed, 44 insertions(+), 8 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/9889ce9d/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index 5984c95..d992425 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -936,9 +936,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/9889ce9d/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java index 9de39dd..9241334 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java @@ -1079,9 +1079,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/9889ce9d/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java -- diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java index be51e96..2d8a74a 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java @@ -246,6 +246,7 @@ public class TestProtobufUtil { scanBuilder.setCacheBlocks(false); scanBuilder.setCaching(1024); scanBuilder.setTimeRange(ProtobufUtil.toTimeRange(TimeRange.allTime())); +scanBuilder.setIncludeStopRow(false); ClientProtos.Scan expectedProto = scanBuilder.build(); ClientProtos.Scan actualProto = ProtobufUtil.toScan( http://git-wip-us.apache.org/repos/asf/hbase/blob/9889ce9d/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java index fad4f45..d55ce1b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java @@ -1090,7 +1090,7 @@ public class TestFromClientSide3 { } Scan scan = new Scan(); -
hbase git commit: HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result
Repository: hbase Updated Branches: refs/heads/branch-2.1 a76a842a6 -> 512a6322d HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/512a6322 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/512a6322 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/512a6322 Branch: refs/heads/branch-2.1 Commit: 512a6322d9a4cf159ccee5a7a01e59bb89967b21 Parents: a76a842 Author: Guanghao Zhang Authored: Thu Dec 20 11:03:54 2018 +0800 Committer: Guanghao Zhang Committed: Fri Dec 21 09:54:58 2018 +0800 -- .../hadoop/hbase/protobuf/ProtobufUtil.java | 4 +-- .../hbase/shaded/protobuf/ProtobufUtil.java | 4 +-- .../hbase/shaded/protobuf/TestProtobufUtil.java | 1 + .../hbase/client/TestFromClientSide3.java | 4 +-- .../client/TestScannersFromClientSide.java | 38 .../hadoop/hbase/protobuf/TestProtobufUtil.java | 1 + 6 files changed, 44 insertions(+), 8 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/512a6322/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index 965aa34..326e78a 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -939,9 +939,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/512a6322/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java index 87eff00..8da0656 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java @@ -1082,9 +1082,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/512a6322/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java -- diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java index be51e96..2d8a74a 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java @@ -246,6 +246,7 @@ public class TestProtobufUtil { scanBuilder.setCacheBlocks(false); scanBuilder.setCaching(1024); scanBuilder.setTimeRange(ProtobufUtil.toTimeRange(TimeRange.allTime())); +scanBuilder.setIncludeStopRow(false); ClientProtos.Scan expectedProto = scanBuilder.build(); ClientProtos.Scan actualProto = ProtobufUtil.toScan( http://git-wip-us.apache.org/repos/asf/hbase/blob/512a6322/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java index fad4f45..d55ce1b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java @@ -1090,7 +1090,7 @@ public class TestFromClientSide3 { } Scan scan = new Scan(); -
hbase git commit: HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result
Repository: hbase Updated Branches: refs/heads/branch-2 f47761f91 -> 1dd137b14 HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/1dd137b1 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/1dd137b1 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/1dd137b1 Branch: refs/heads/branch-2 Commit: 1dd137b148a68046815b0137917a58c3cf2050f3 Parents: f47761f Author: Guanghao Zhang Authored: Thu Dec 20 11:03:54 2018 +0800 Committer: Guanghao Zhang Committed: Fri Dec 21 09:50:23 2018 +0800 -- .../hadoop/hbase/protobuf/ProtobufUtil.java | 4 +-- .../hbase/shaded/protobuf/ProtobufUtil.java | 4 +-- .../hbase/shaded/protobuf/TestProtobufUtil.java | 1 + .../hbase/client/TestFromClientSide3.java | 4 +-- .../client/TestScannersFromClientSide.java | 38 .../hadoop/hbase/protobuf/TestProtobufUtil.java | 1 + 6 files changed, 44 insertions(+), 8 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/1dd137b1/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index 965aa34..326e78a 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -939,9 +939,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/1dd137b1/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java index ee60d87..3039392 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java @@ -1082,9 +1082,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/1dd137b1/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java -- diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java index be51e96..2d8a74a 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java @@ -246,6 +246,7 @@ public class TestProtobufUtil { scanBuilder.setCacheBlocks(false); scanBuilder.setCaching(1024); scanBuilder.setTimeRange(ProtobufUtil.toTimeRange(TimeRange.allTime())); +scanBuilder.setIncludeStopRow(false); ClientProtos.Scan expectedProto = scanBuilder.build(); ClientProtos.Scan actualProto = ProtobufUtil.toScan( http://git-wip-us.apache.org/repos/asf/hbase/blob/1dd137b1/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java index fad4f45..d55ce1b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java @@ -1090,7 +1090,7 @@ public class TestFromClientSide3 { } Scan scan = new Scan(); -
hbase git commit: HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result
Repository: hbase Updated Branches: refs/heads/master 787567336 -> ad819380c HBASE-21618 Scan with the same startRow(inclusive=true) and stopRow(inclusive=false) returns one result Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/ad819380 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/ad819380 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/ad819380 Branch: refs/heads/master Commit: ad819380c744678e719431fb8b1b5e1951bc31b6 Parents: 7875673 Author: Guanghao Zhang Authored: Thu Dec 20 11:03:54 2018 +0800 Committer: Guanghao Zhang Committed: Fri Dec 21 09:49:24 2018 +0800 -- .../hadoop/hbase/protobuf/ProtobufUtil.java | 4 +-- .../hbase/shaded/protobuf/ProtobufUtil.java | 4 +-- .../hbase/shaded/protobuf/TestProtobufUtil.java | 1 + .../hbase/client/TestFromClientSide3.java | 4 +-- .../client/TestScannersFromClientSide.java | 38 .../hadoop/hbase/protobuf/TestProtobufUtil.java | 1 + 6 files changed, 44 insertions(+), 8 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/ad819380/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index 4d54528..a3d49b5 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -937,9 +937,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/ad819380/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java index cf4c831..fea81f1 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java @@ -1081,9 +1081,7 @@ public final class ProtobufUtil { if (!scan.includeStartRow()) { scanBuilder.setIncludeStartRow(false); } -if (scan.includeStopRow()) { - scanBuilder.setIncludeStopRow(true); -} +scanBuilder.setIncludeStopRow(scan.includeStopRow()); if (scan.getReadType() != Scan.ReadType.DEFAULT) { scanBuilder.setReadType(toReadType(scan.getReadType())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/ad819380/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java -- diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java index be51e96..2d8a74a 100644 --- a/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java +++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/shaded/protobuf/TestProtobufUtil.java @@ -246,6 +246,7 @@ public class TestProtobufUtil { scanBuilder.setCacheBlocks(false); scanBuilder.setCaching(1024); scanBuilder.setTimeRange(ProtobufUtil.toTimeRange(TimeRange.allTime())); +scanBuilder.setIncludeStopRow(false); ClientProtos.Scan expectedProto = scanBuilder.build(); ClientProtos.Scan actualProto = ProtobufUtil.toScan( http://git-wip-us.apache.org/repos/asf/hbase/blob/ad819380/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java index 0dee20b..cbfa1bf 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide3.java @@ -1094,7 +1094,7 @@ public class TestFromClientSide3 { } Scan scan = new Scan(); -
hbase git commit: HBASE-21610, numOpenConnections metric is set to -1 when zero server channel exist
Repository: hbase Updated Branches: refs/heads/branch-2 83e12153d -> f47761f91 HBASE-21610, numOpenConnections metric is set to -1 when zero server channel exist Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/f47761f9 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/f47761f9 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/f47761f9 Branch: refs/heads/branch-2 Commit: f47761f91c9eb33e747d4dd40866e580d915dca5 Parents: 83e1215 Author: Pankaj Authored: Tue Dec 18 01:31:55 2018 +0530 Committer: stack Committed: Thu Dec 20 16:36:23 2018 -0800 -- .../src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/f47761f9/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java index 8ea2057..742a728 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java @@ -169,8 +169,9 @@ public class NettyRpcServer extends RpcServer { @Override public int getNumOpenConnections() { +int channelsCount = allChannels.size(); // allChannels also contains the server channel, so exclude that from the count. -return allChannels.size() - 1; +return channelsCount > 0 ? channelsCount - 1 : channelsCount; } @Override
hbase git commit: HBASE-21610, numOpenConnections metric is set to -1 when zero server channel exist
Repository: hbase Updated Branches: refs/heads/master 8991877bb -> 787567336 HBASE-21610, numOpenConnections metric is set to -1 when zero server channel exist Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/78756733 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/78756733 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/78756733 Branch: refs/heads/master Commit: 787567336afb9c5c1e00aaa0326566a5522a5e31 Parents: 8991877 Author: Pankaj Authored: Tue Dec 18 01:31:55 2018 +0530 Committer: stack Committed: Thu Dec 20 16:36:42 2018 -0800 -- .../src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/78756733/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java index 8ea2057..742a728 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java @@ -169,8 +169,9 @@ public class NettyRpcServer extends RpcServer { @Override public int getNumOpenConnections() { +int channelsCount = allChannels.size(); // allChannels also contains the server channel, so exclude that from the count. -return allChannels.size() - 1; +return channelsCount > 0 ? channelsCount - 1 : channelsCount; } @Override
hbase git commit: HBASE-21610, numOpenConnections metric is set to -1 when zero server channel exist
Repository: hbase Updated Branches: refs/heads/branch-2.1 27a0f205c -> a76a842a6 HBASE-21610, numOpenConnections metric is set to -1 when zero server channel exist Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/a76a842a Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/a76a842a Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/a76a842a Branch: refs/heads/branch-2.1 Commit: a76a842a619ccc90dcdfe52e2b690a8d5ab92035 Parents: 27a0f20 Author: Pankaj Authored: Tue Dec 18 01:31:55 2018 +0530 Committer: stack Committed: Thu Dec 20 16:36:02 2018 -0800 -- .../src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/a76a842a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java index 8ea2057..742a728 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java @@ -169,8 +169,9 @@ public class NettyRpcServer extends RpcServer { @Override public int getNumOpenConnections() { +int channelsCount = allChannels.size(); // allChannels also contains the server channel, so exclude that from the count. -return allChannels.size() - 1; +return channelsCount > 0 ? channelsCount - 1 : channelsCount; } @Override
hbase git commit: HBASE-21610, numOpenConnections metric is set to -1 when zero server channel exist
Repository: hbase Updated Branches: refs/heads/branch-2.0 b07a28c56 -> 7e51e781f HBASE-21610, numOpenConnections metric is set to -1 when zero server channel exist Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/7e51e781 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/7e51e781 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/7e51e781 Branch: refs/heads/branch-2.0 Commit: 7e51e781f376a5229b9d4638724153840feec981 Parents: b07a28c Author: Pankaj Authored: Tue Dec 18 01:31:55 2018 +0530 Committer: stack Committed: Thu Dec 20 16:35:34 2018 -0800 -- .../src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/7e51e781/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java index 8ea2057..742a728 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java @@ -169,8 +169,9 @@ public class NettyRpcServer extends RpcServer { @Override public int getNumOpenConnections() { +int channelsCount = allChannels.size(); // allChannels also contains the server channel, so exclude that from the count. -return allChannels.size() - 1; +return channelsCount > 0 ? channelsCount - 1 : channelsCount; } @Override
[8/9] hbase-site git commit: Published site at 8991877bb250ee1fe66c2b9a491645973927d674.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html index 3196572..659f14c 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html @@ -2494,7 +2494,7 @@ implements waitForMetaOnline -publicbooleanwaitForMetaOnline() +publicbooleanwaitForMetaOnline() throws https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException Check hbase:meta is up and ready for reading. For use during Master startup only. @@ -2512,7 +2512,7 @@ implements isRegionOnline -privatebooleanisRegionOnline(RegionInfori) +privatebooleanisRegionOnline(RegionInfori) throws https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException Returns: @@ -2529,7 +2529,7 @@ implements waitForNamespaceOnline -privatebooleanwaitForNamespaceOnline() +privatebooleanwaitForNamespaceOnline() throws https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException, https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Check hbase:namespace table is assigned. If not, startup will hang looking for the ns table @@ -2552,7 +2552,7 @@ implements updateConfigurationForQuotasObserver -publicvoidupdateConfigurationForQuotasObserver(org.apache.hadoop.conf.Configurationconf) +publicvoidupdateConfigurationForQuotasObserver(org.apache.hadoop.conf.Configurationconf) Adds the MasterQuotasObserver to the list of configured Master observers to automatically remove quotas for a table when that table is deleted. @@ -2563,7 +2563,7 @@ implements initMobCleaner -privatevoidinitMobCleaner() +privatevoidinitMobCleaner() @@ -2572,7 +2572,7 @@ implements createMetaBootstrap -protectedMasterMetaBootstrapcreateMetaBootstrap() +protectedMasterMetaBootstrapcreateMetaBootstrap() Create a MasterMetaBootstrap instance. @@ -2587,7 +2587,7 @@ implements createServerManager -protectedServerManagercreateServerManager(MasterServicesmaster) +protectedServerManagercreateServerManager(MasterServicesmaster) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Create a ServerManager instance. @@ -2607,7 +2607,7 @@ implements waitForRegionServers -privatevoidwaitForRegionServers(MonitoredTaskstatus) +privatevoidwaitForRegionServers(MonitoredTaskstatus) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException, https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException @@ -2623,7 +2623,7 @@ implements initClusterSchemaService -protectedvoidinitClusterSchemaService() +protectedvoidinitClusterSchemaService() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException, https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException @@ -2639,7 +2639,7 @@ implements initQuotaManager -privatevoidinitQuotaManager() +privatevoidinitQuotaManager() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -2653,7 +2653,7 @@ implements createQuotaSnapshotNotifier -privateSpaceQuotaSnapshotNotifiercreateQuotaSnapshotNotifier() +privateSpaceQuotaSnapshotNotifiercreateQuotaSnapshotNotifier() @@ -2662,7 +2662,7 @@ implements isCatalogJanitorEnabled -booleanisCatalogJanitorEnabled() +booleanisCatalogJanitorEnabled() @@ -2671,7 +2671,7 @@ implements isCleanerChoreEnabled -booleanisCleanerChoreEnabled() +booleanisCleanerChoreEnabled() @@ -2680,7 +2680,7 @@ implements getServerManager -publicServerManagergetServerManager() +publicServerManagergetServerManager() Specified by: getServerManagerin interfaceMasterServices @@ -2695,7 +2695,7 @@ implements
[9/9] hbase-site git commit: Published site at 8991877bb250ee1fe66c2b9a491645973927d674.
Published site at 8991877bb250ee1fe66c2b9a491645973927d674. Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/1825b69e Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/1825b69e Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/1825b69e Branch: refs/heads/asf-site Commit: 1825b69e504db8c9a3e570f17a2f91a345be530d Parents: a767800 Author: jenkins Authored: Thu Dec 20 10:50:31 2018 + Committer: jenkins Committed: Thu Dec 20 10:50:31 2018 + -- acid-semantics.html |4 +- apache_hbase_reference_guide.pdf|4 +- book.html |2 +- bulk-loads.html |4 +- checkstyle-aggregate.html | 254 +- coc.html|4 +- dependencies.html |4 +- dependency-convergence.html |4 +- dependency-info.html|4 +- dependency-management.html |4 +- devapidocs/constant-values.html |4 +- .../master/HMaster.MasterStoppedException.html |4 +- .../master/HMaster.TableDescriptorGetter.html |4 +- .../org/apache/hadoop/hbase/master/HMaster.html | 316 +- .../org/apache/hadoop/hbase/Version.html|4 +- ...ReaderImpl.BlockIndexNotLoadedException.html |4 +- .../hfile/HFileReaderImpl.EncodedScanner.html |4 +- .../hfile/HFileReaderImpl.HFileScannerImpl.html |4 +- .../HFileReaderImpl.NotSeekedException.html |4 +- .../hadoop/hbase/io/hfile/HFileReaderImpl.html |4 +- .../master/HMaster.InitializationMonitor.html | 6063 +- .../master/HMaster.MasterStoppedException.html | 6063 +- .../hbase/master/HMaster.RedirectServlet.html | 6063 +- .../master/HMaster.TableDescriptorGetter.html | 6063 +- .../org/apache/hadoop/hbase/master/HMaster.html | 6063 +- downloads.html |4 +- export_control.html |4 +- index.html |4 +- integration.html|4 +- issue-tracking.html |4 +- license.html|4 +- mail-lists.html |4 +- metrics.html|4 +- old_news.html |4 +- plugin-management.html |4 +- plugins.html|4 +- poweredbyhbase.html |4 +- project-info.html |4 +- project-reports.html|4 +- project-summary.html|4 +- pseudo-distributed.html |4 +- replication.html|4 +- resources.html |4 +- source-repository.html |4 +- sponsors.html |4 +- supportingprojects.html |4 +- team-list.html |4 +- 47 files changed, 15524 insertions(+), 15519 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/acid-semantics.html -- diff --git a/acid-semantics.html b/acid-semantics.html index da6bdeb..44bf748 100644 --- a/acid-semantics.html +++ b/acid-semantics.html @@ -7,7 +7,7 @@ - + Apache HBase Apache HBase (TM) ACID Properties @@ -611,7 +611,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-12-18 + Last Published: 2018-12-20 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/apache_hbase_reference_guide.pdf -- diff --git a/apache_hbase_reference_guide.pdf b/apache_hbase_reference_guide.pdf index 61a72b1..5e4f290 100644 --- a/apache_hbase_reference_guide.pdf +++ b/apache_hbase_reference_guide.pdf @@ -5,8 +5,8 @@ /Author (Apache HBase Team) /Creator (Asciidoctor PDF 1.5.0.alpha.15, based on Prawn 2.2.2) /Producer (Apache HBase Team) -/ModDate (D:20181218143412+00'00') -/CreationDate (D:20181218145148+00'00') +/ModDate (D:20181220102940+00'00') +/CreationDate (D:20181220104643+00'00') >> endobj
hbase-site git commit: INFRA-10751 Empty commit
Repository: hbase-site Updated Branches: refs/heads/asf-site 1825b69e5 -> 54270554e INFRA-10751 Empty commit Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/54270554 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/54270554 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/54270554 Branch: refs/heads/asf-site Commit: 54270554e402aee57f4c09d26cba6b1831aae253 Parents: 1825b69 Author: jenkins Authored: Thu Dec 20 10:50:43 2018 + Committer: jenkins Committed: Thu Dec 20 10:50:43 2018 + -- --
[3/9] hbase-site git commit: Published site at 8991877bb250ee1fe66c2b9a491645973927d674.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html index fd8ff3b..f14e153 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html @@ -908,3038 +908,3039 @@ 900 */ 901 private void finishActiveMasterInitialization(MonitoredTask status) throws IOException, 902 InterruptedException, KeeperException, ReplicationException { -903Thread zombieDetector = new Thread(new InitializationMonitor(this), -904 "ActiveMasterInitializationMonitor-" + System.currentTimeMillis()); -905zombieDetector.setDaemon(true); -906zombieDetector.start(); +903/* +904 * We are active master now... go initialize components we need to run. +905 */ +906status.setStatus("Initializing Master file system"); 907 -908/* -909 * We are active master now... go initialize components we need to run. -910 */ -911status.setStatus("Initializing Master file system"); -912 -913this.masterActiveTime = System.currentTimeMillis(); -914// TODO: Do this using Dependency Injection, using PicoContainer, Guice or Spring. -915 -916// Only initialize the MemStoreLAB when master carry table -917if (LoadBalancer.isTablesOnMaster(conf)) { -918 initializeMemStoreChunkCreator(); -919} -920this.fileSystemManager = new MasterFileSystem(conf); -921this.walManager = new MasterWalManager(this); -922 -923// enable table descriptors cache -924this.tableDescriptors.setCacheOn(); -925 -926// warm-up HTDs cache on master initialization -927if (preLoadTableDescriptors) { -928 status.setStatus("Pre-loading table descriptors"); -929 this.tableDescriptors.getAll(); -930} -931 -932// Publish cluster ID; set it in Master too. The superclass RegionServer does this later but -933// only after it has checked in with the Master. At least a few tests ask Master for clusterId -934// before it has called its run method and before RegionServer has done the reportForDuty. -935ClusterId clusterId = fileSystemManager.getClusterId(); -936status.setStatus("Publishing Cluster ID " + clusterId + " in ZooKeeper"); -937 ZKClusterId.setClusterId(this.zooKeeper, fileSystemManager.getClusterId()); -938this.clusterId = clusterId.toString(); -939 -940// Precaution. Put in place the old hbck1 lock file to fence out old hbase1s running their -941// hbck1s against an hbase2 cluster; it could do damage. To skip this behavior, set -942// hbase.write.hbck1.lock.file to false. -943if (this.conf.getBoolean("hbase.write.hbck1.lock.file", true)) { -944 HBaseFsck.checkAndMarkRunningHbck(this.conf, -945 HBaseFsck.createLockRetryCounterFactory(this.conf).create()); -946} -947 -948status.setStatus("Initialize ServerManager and schedule SCP for crash servers"); -949// The below two managers must be created before loading procedures, as they will be used during -950// loading. -951this.serverManager = createServerManager(this); -952this.syncReplicationReplayWALManager = new SyncReplicationReplayWALManager(this); -953createProcedureExecutor(); -954@SuppressWarnings("rawtypes") -955MapClass? extends Procedure, ListProcedureMasterProcedureEnv procsByType = -956 procedureExecutor.getActiveProceduresNoCopy().stream() -957.collect(Collectors.groupingBy(p - p.getClass())); -958 -959 checkUnsupportedProcedure(procsByType); -960 -961// Create Assignment Manager -962this.assignmentManager = createAssignmentManager(this); -963this.assignmentManager.start(); -964// TODO: TRSP can perform as the sub procedure for other procedures, so even if it is marked as -965// completed, it could still be in the procedure list. This is a bit strange but is another -966// story, need to verify the implementation for ProcedureExecutor and ProcedureStore. -967 ListTransitRegionStateProcedure ritList = -968 procsByType.getOrDefault(TransitRegionStateProcedure.class, Collections.emptyList()).stream() -969.filter(p - !p.isFinished()).map(p - (TransitRegionStateProcedure) p) -970.collect(Collectors.toList()); -971 this.assignmentManager.setupRIT(ritList); -972 -973// Start RegionServerTracker with listing of servers found with exiting SCPs -- these should -974// be registered in the deadServers set -- and with the list of servernames out on the -975// filesystem that COULD BE 'alive' (we'll schedule SCPs for each and let
[2/9] hbase-site git commit: Published site at 8991877bb250ee1fe66c2b9a491645973927d674.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html index fd8ff3b..f14e153 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html @@ -908,3038 +908,3039 @@ 900 */ 901 private void finishActiveMasterInitialization(MonitoredTask status) throws IOException, 902 InterruptedException, KeeperException, ReplicationException { -903Thread zombieDetector = new Thread(new InitializationMonitor(this), -904 "ActiveMasterInitializationMonitor-" + System.currentTimeMillis()); -905zombieDetector.setDaemon(true); -906zombieDetector.start(); +903/* +904 * We are active master now... go initialize components we need to run. +905 */ +906status.setStatus("Initializing Master file system"); 907 -908/* -909 * We are active master now... go initialize components we need to run. -910 */ -911status.setStatus("Initializing Master file system"); -912 -913this.masterActiveTime = System.currentTimeMillis(); -914// TODO: Do this using Dependency Injection, using PicoContainer, Guice or Spring. -915 -916// Only initialize the MemStoreLAB when master carry table -917if (LoadBalancer.isTablesOnMaster(conf)) { -918 initializeMemStoreChunkCreator(); -919} -920this.fileSystemManager = new MasterFileSystem(conf); -921this.walManager = new MasterWalManager(this); -922 -923// enable table descriptors cache -924this.tableDescriptors.setCacheOn(); -925 -926// warm-up HTDs cache on master initialization -927if (preLoadTableDescriptors) { -928 status.setStatus("Pre-loading table descriptors"); -929 this.tableDescriptors.getAll(); -930} -931 -932// Publish cluster ID; set it in Master too. The superclass RegionServer does this later but -933// only after it has checked in with the Master. At least a few tests ask Master for clusterId -934// before it has called its run method and before RegionServer has done the reportForDuty. -935ClusterId clusterId = fileSystemManager.getClusterId(); -936status.setStatus("Publishing Cluster ID " + clusterId + " in ZooKeeper"); -937 ZKClusterId.setClusterId(this.zooKeeper, fileSystemManager.getClusterId()); -938this.clusterId = clusterId.toString(); -939 -940// Precaution. Put in place the old hbck1 lock file to fence out old hbase1s running their -941// hbck1s against an hbase2 cluster; it could do damage. To skip this behavior, set -942// hbase.write.hbck1.lock.file to false. -943if (this.conf.getBoolean("hbase.write.hbck1.lock.file", true)) { -944 HBaseFsck.checkAndMarkRunningHbck(this.conf, -945 HBaseFsck.createLockRetryCounterFactory(this.conf).create()); -946} -947 -948status.setStatus("Initialize ServerManager and schedule SCP for crash servers"); -949// The below two managers must be created before loading procedures, as they will be used during -950// loading. -951this.serverManager = createServerManager(this); -952this.syncReplicationReplayWALManager = new SyncReplicationReplayWALManager(this); -953createProcedureExecutor(); -954@SuppressWarnings("rawtypes") -955MapClass? extends Procedure, ListProcedureMasterProcedureEnv procsByType = -956 procedureExecutor.getActiveProceduresNoCopy().stream() -957.collect(Collectors.groupingBy(p - p.getClass())); -958 -959 checkUnsupportedProcedure(procsByType); -960 -961// Create Assignment Manager -962this.assignmentManager = createAssignmentManager(this); -963this.assignmentManager.start(); -964// TODO: TRSP can perform as the sub procedure for other procedures, so even if it is marked as -965// completed, it could still be in the procedure list. This is a bit strange but is another -966// story, need to verify the implementation for ProcedureExecutor and ProcedureStore. -967 ListTransitRegionStateProcedure ritList = -968 procsByType.getOrDefault(TransitRegionStateProcedure.class, Collections.emptyList()).stream() -969.filter(p - !p.isFinished()).map(p - (TransitRegionStateProcedure) p) -970.collect(Collectors.toList()); -971 this.assignmentManager.setupRIT(ritList); -972 -973// Start RegionServerTracker with listing of servers found with exiting SCPs -- these should -974// be registered in the deadServers set -- and with the list of servernames out on the -975// filesystem that COULD BE 'alive' (we'll schedule SCPs for each and let SCP figure it out). -976// We also pass dirs that are already 'splitting'... so we can do some checks down in
[1/9] hbase-site git commit: Published site at 8991877bb250ee1fe66c2b9a491645973927d674.
Repository: hbase-site Updated Branches: refs/heads/asf-site a76780052 -> 1825b69e5 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/downloads.html -- diff --git a/downloads.html b/downloads.html index afec59a..3eee98b 100644 --- a/downloads.html +++ b/downloads.html @@ -7,7 +7,7 @@ - + Apache HBase Apache HBase Downloads @@ -463,7 +463,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-12-18 + Last Published: 2018-12-20 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/export_control.html -- diff --git a/export_control.html b/export_control.html index 8f10dd1..716b079 100644 --- a/export_control.html +++ b/export_control.html @@ -7,7 +7,7 @@ - + Apache HBase Export Control @@ -341,7 +341,7 @@ for more details. https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-12-18 + Last Published: 2018-12-20 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/index.html -- diff --git a/index.html b/index.html index 2f11423..2657cbd 100644 --- a/index.html +++ b/index.html @@ -7,7 +7,7 @@ - + Apache HBase Apache HBase⢠Home @@ -421,7 +421,7 @@ Apache HBase is an open-source, distributed, versioned, non-relational database https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-12-18 + Last Published: 2018-12-20 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/integration.html -- diff --git a/integration.html b/integration.html index a9ad220..b88951c 100644 --- a/integration.html +++ b/integration.html @@ -7,7 +7,7 @@ - + Apache HBase CI Management @@ -301,7 +301,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-12-18 + Last Published: 2018-12-20 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/issue-tracking.html -- diff --git a/issue-tracking.html b/issue-tracking.html index c0963ab..e70228a 100644 --- a/issue-tracking.html +++ b/issue-tracking.html @@ -7,7 +7,7 @@ - + Apache HBase Issue Management @@ -298,7 +298,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-12-18 + Last Published: 2018-12-20 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/license.html -- diff --git a/license.html b/license.html index ec99e9b..2c01a72 100644 --- a/license.html +++ b/license.html @@ -7,7 +7,7 @@ - + Apache HBase Project Licenses @@ -501,7 +501,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-12-18 + Last Published: 2018-12-20 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/mail-lists.html -- diff --git a/mail-lists.html b/mail-lists.html index 5b106bd..234a8dd 100644 --- a/mail-lists.html +++ b/mail-lists.html @@ -7,7 +7,7 @@ - + Apache HBase Project Mailing Lists @@ -351,7 +351,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-12-18 + Last Published: 2018-12-20 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/metrics.html -- diff --git
[5/9] hbase-site git commit: Published site at 8991877bb250ee1fe66c2b9a491645973927d674.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html index fd8ff3b..f14e153 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html @@ -908,3038 +908,3039 @@ 900 */ 901 private void finishActiveMasterInitialization(MonitoredTask status) throws IOException, 902 InterruptedException, KeeperException, ReplicationException { -903Thread zombieDetector = new Thread(new InitializationMonitor(this), -904 "ActiveMasterInitializationMonitor-" + System.currentTimeMillis()); -905zombieDetector.setDaemon(true); -906zombieDetector.start(); +903/* +904 * We are active master now... go initialize components we need to run. +905 */ +906status.setStatus("Initializing Master file system"); 907 -908/* -909 * We are active master now... go initialize components we need to run. -910 */ -911status.setStatus("Initializing Master file system"); -912 -913this.masterActiveTime = System.currentTimeMillis(); -914// TODO: Do this using Dependency Injection, using PicoContainer, Guice or Spring. -915 -916// Only initialize the MemStoreLAB when master carry table -917if (LoadBalancer.isTablesOnMaster(conf)) { -918 initializeMemStoreChunkCreator(); -919} -920this.fileSystemManager = new MasterFileSystem(conf); -921this.walManager = new MasterWalManager(this); -922 -923// enable table descriptors cache -924this.tableDescriptors.setCacheOn(); -925 -926// warm-up HTDs cache on master initialization -927if (preLoadTableDescriptors) { -928 status.setStatus("Pre-loading table descriptors"); -929 this.tableDescriptors.getAll(); -930} -931 -932// Publish cluster ID; set it in Master too. The superclass RegionServer does this later but -933// only after it has checked in with the Master. At least a few tests ask Master for clusterId -934// before it has called its run method and before RegionServer has done the reportForDuty. -935ClusterId clusterId = fileSystemManager.getClusterId(); -936status.setStatus("Publishing Cluster ID " + clusterId + " in ZooKeeper"); -937 ZKClusterId.setClusterId(this.zooKeeper, fileSystemManager.getClusterId()); -938this.clusterId = clusterId.toString(); -939 -940// Precaution. Put in place the old hbck1 lock file to fence out old hbase1s running their -941// hbck1s against an hbase2 cluster; it could do damage. To skip this behavior, set -942// hbase.write.hbck1.lock.file to false. -943if (this.conf.getBoolean("hbase.write.hbck1.lock.file", true)) { -944 HBaseFsck.checkAndMarkRunningHbck(this.conf, -945 HBaseFsck.createLockRetryCounterFactory(this.conf).create()); -946} -947 -948status.setStatus("Initialize ServerManager and schedule SCP for crash servers"); -949// The below two managers must be created before loading procedures, as they will be used during -950// loading. -951this.serverManager = createServerManager(this); -952this.syncReplicationReplayWALManager = new SyncReplicationReplayWALManager(this); -953createProcedureExecutor(); -954@SuppressWarnings("rawtypes") -955MapClass? extends Procedure, ListProcedureMasterProcedureEnv procsByType = -956 procedureExecutor.getActiveProceduresNoCopy().stream() -957.collect(Collectors.groupingBy(p - p.getClass())); -958 -959 checkUnsupportedProcedure(procsByType); -960 -961// Create Assignment Manager -962this.assignmentManager = createAssignmentManager(this); -963this.assignmentManager.start(); -964// TODO: TRSP can perform as the sub procedure for other procedures, so even if it is marked as -965// completed, it could still be in the procedure list. This is a bit strange but is another -966// story, need to verify the implementation for ProcedureExecutor and ProcedureStore. -967 ListTransitRegionStateProcedure ritList = -968 procsByType.getOrDefault(TransitRegionStateProcedure.class, Collections.emptyList()).stream() -969.filter(p - !p.isFinished()).map(p - (TransitRegionStateProcedure) p) -970.collect(Collectors.toList()); -971 this.assignmentManager.setupRIT(ritList); -972 -973// Start RegionServerTracker with listing of servers found with exiting SCPs -- these should -974// be registered in the deadServers set -- and with the list of servernames out on the -975// filesystem that COULD BE 'alive' (we'll schedule SCPs for each and
[4/9] hbase-site git commit: Published site at 8991877bb250ee1fe66c2b9a491645973927d674.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html index fd8ff3b..f14e153 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html @@ -908,3038 +908,3039 @@ 900 */ 901 private void finishActiveMasterInitialization(MonitoredTask status) throws IOException, 902 InterruptedException, KeeperException, ReplicationException { -903Thread zombieDetector = new Thread(new InitializationMonitor(this), -904 "ActiveMasterInitializationMonitor-" + System.currentTimeMillis()); -905zombieDetector.setDaemon(true); -906zombieDetector.start(); +903/* +904 * We are active master now... go initialize components we need to run. +905 */ +906status.setStatus("Initializing Master file system"); 907 -908/* -909 * We are active master now... go initialize components we need to run. -910 */ -911status.setStatus("Initializing Master file system"); -912 -913this.masterActiveTime = System.currentTimeMillis(); -914// TODO: Do this using Dependency Injection, using PicoContainer, Guice or Spring. -915 -916// Only initialize the MemStoreLAB when master carry table -917if (LoadBalancer.isTablesOnMaster(conf)) { -918 initializeMemStoreChunkCreator(); -919} -920this.fileSystemManager = new MasterFileSystem(conf); -921this.walManager = new MasterWalManager(this); -922 -923// enable table descriptors cache -924this.tableDescriptors.setCacheOn(); -925 -926// warm-up HTDs cache on master initialization -927if (preLoadTableDescriptors) { -928 status.setStatus("Pre-loading table descriptors"); -929 this.tableDescriptors.getAll(); -930} -931 -932// Publish cluster ID; set it in Master too. The superclass RegionServer does this later but -933// only after it has checked in with the Master. At least a few tests ask Master for clusterId -934// before it has called its run method and before RegionServer has done the reportForDuty. -935ClusterId clusterId = fileSystemManager.getClusterId(); -936status.setStatus("Publishing Cluster ID " + clusterId + " in ZooKeeper"); -937 ZKClusterId.setClusterId(this.zooKeeper, fileSystemManager.getClusterId()); -938this.clusterId = clusterId.toString(); -939 -940// Precaution. Put in place the old hbck1 lock file to fence out old hbase1s running their -941// hbck1s against an hbase2 cluster; it could do damage. To skip this behavior, set -942// hbase.write.hbck1.lock.file to false. -943if (this.conf.getBoolean("hbase.write.hbck1.lock.file", true)) { -944 HBaseFsck.checkAndMarkRunningHbck(this.conf, -945 HBaseFsck.createLockRetryCounterFactory(this.conf).create()); -946} -947 -948status.setStatus("Initialize ServerManager and schedule SCP for crash servers"); -949// The below two managers must be created before loading procedures, as they will be used during -950// loading. -951this.serverManager = createServerManager(this); -952this.syncReplicationReplayWALManager = new SyncReplicationReplayWALManager(this); -953createProcedureExecutor(); -954@SuppressWarnings("rawtypes") -955MapClass? extends Procedure, ListProcedureMasterProcedureEnv procsByType = -956 procedureExecutor.getActiveProceduresNoCopy().stream() -957.collect(Collectors.groupingBy(p - p.getClass())); -958 -959 checkUnsupportedProcedure(procsByType); -960 -961// Create Assignment Manager -962this.assignmentManager = createAssignmentManager(this); -963this.assignmentManager.start(); -964// TODO: TRSP can perform as the sub procedure for other procedures, so even if it is marked as -965// completed, it could still be in the procedure list. This is a bit strange but is another -966// story, need to verify the implementation for ProcedureExecutor and ProcedureStore. -967 ListTransitRegionStateProcedure ritList = -968 procsByType.getOrDefault(TransitRegionStateProcedure.class, Collections.emptyList()).stream() -969.filter(p - !p.isFinished()).map(p - (TransitRegionStateProcedure) p) -970.collect(Collectors.toList()); -971 this.assignmentManager.setupRIT(ritList); -972 -973// Start RegionServerTracker with listing of servers found with exiting SCPs -- these should -974// be registered in the deadServers set -- and with the list of servernames out on the -975// filesystem that COULD BE 'alive' (we'll schedule SCPs for each and let SCP figure it out). -976//
[7/9] hbase-site git commit: Published site at 8991877bb250ee1fe66c2b9a491645973927d674.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/src-html/org/apache/hadoop/hbase/Version.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html index 5a2a23d..c579d19 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html @@ -18,9 +18,9 @@ 010 public static final String version = "3.0.0-SNAPSHOT"; 011 public static final String revision = ""; 012 public static final String user = "jenkins"; -013 public static final String date = "Tue Dec 18 14:46:13 UTC 2018"; +013 public static final String date = "Thu Dec 20 10:41:19 UTC 2018"; 014 public static final String url = "git://jenkins-websites1.apache.org/home/jenkins/jenkins-slave/workspace/hbase_generate_website/hbase"; -015 public static final String srcChecksum = "0ff9598693406383f0355b09c8c7ea64"; +015 public static final String srcChecksum = "cac6c13f455cde23001f256bae768095"; 016} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.BlockIndexNotLoadedException.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.BlockIndexNotLoadedException.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.BlockIndexNotLoadedException.html index d8d391b..ab910a7 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.BlockIndexNotLoadedException.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.BlockIndexNotLoadedException.html @@ -1308,8 +1308,8 @@ 1300 boolean isCompaction, boolean updateCacheMetrics, BlockType expectedBlockType, 1301 DataBlockEncoding expectedDataBlockEncoding) throws IOException { 1302// Check cache for block. If found return. -1303if (cacheConf.getBlockCache().isPresent()) { -1304 BlockCache cache = cacheConf.getBlockCache().get(); +1303BlockCache cache = cacheConf.getBlockCache().orElse(null); +1304if (cache != null) { 1305 HFileBlock cachedBlock = 1306 (HFileBlock) cache.getBlock(cacheKey, cacheBlock, useLock, updateCacheMetrics); 1307 if (cachedBlock != null) { http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html index d8d391b..ab910a7 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html @@ -1308,8 +1308,8 @@ 1300 boolean isCompaction, boolean updateCacheMetrics, BlockType expectedBlockType, 1301 DataBlockEncoding expectedDataBlockEncoding) throws IOException { 1302// Check cache for block. If found return. -1303if (cacheConf.getBlockCache().isPresent()) { -1304 BlockCache cache = cacheConf.getBlockCache().get(); +1303BlockCache cache = cacheConf.getBlockCache().orElse(null); +1304if (cache != null) { 1305 HFileBlock cachedBlock = 1306 (HFileBlock) cache.getBlock(cacheKey, cacheBlock, useLock, updateCacheMetrics); 1307 if (cachedBlock != null) { http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.HFileScannerImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.HFileScannerImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.HFileScannerImpl.html index d8d391b..ab910a7 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.HFileScannerImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.HFileScannerImpl.html @@ -1308,8 +1308,8 @@ 1300 boolean isCompaction, boolean updateCacheMetrics, BlockType expectedBlockType, 1301 DataBlockEncoding expectedDataBlockEncoding) throws IOException { 1302// Check cache for block. If found return. -1303if (cacheConf.getBlockCache().isPresent()) { -1304 BlockCache cache = cacheConf.getBlockCache().get(); +1303BlockCache cache = cacheConf.getBlockCache().orElse(null); +1304if (cache != null) { 1305 HFileBlock cachedBlock = 1306 (HFileBlock) cache.getBlock(cacheKey, cacheBlock, useLock, updateCacheMetrics);
[6/9] hbase-site git commit: Published site at 8991877bb250ee1fe66c2b9a491645973927d674.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1825b69e/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html index fd8ff3b..f14e153 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html @@ -908,3038 +908,3039 @@ 900 */ 901 private void finishActiveMasterInitialization(MonitoredTask status) throws IOException, 902 InterruptedException, KeeperException, ReplicationException { -903Thread zombieDetector = new Thread(new InitializationMonitor(this), -904 "ActiveMasterInitializationMonitor-" + System.currentTimeMillis()); -905zombieDetector.setDaemon(true); -906zombieDetector.start(); +903/* +904 * We are active master now... go initialize components we need to run. +905 */ +906status.setStatus("Initializing Master file system"); 907 -908/* -909 * We are active master now... go initialize components we need to run. -910 */ -911status.setStatus("Initializing Master file system"); -912 -913this.masterActiveTime = System.currentTimeMillis(); -914// TODO: Do this using Dependency Injection, using PicoContainer, Guice or Spring. -915 -916// Only initialize the MemStoreLAB when master carry table -917if (LoadBalancer.isTablesOnMaster(conf)) { -918 initializeMemStoreChunkCreator(); -919} -920this.fileSystemManager = new MasterFileSystem(conf); -921this.walManager = new MasterWalManager(this); -922 -923// enable table descriptors cache -924this.tableDescriptors.setCacheOn(); -925 -926// warm-up HTDs cache on master initialization -927if (preLoadTableDescriptors) { -928 status.setStatus("Pre-loading table descriptors"); -929 this.tableDescriptors.getAll(); -930} -931 -932// Publish cluster ID; set it in Master too. The superclass RegionServer does this later but -933// only after it has checked in with the Master. At least a few tests ask Master for clusterId -934// before it has called its run method and before RegionServer has done the reportForDuty. -935ClusterId clusterId = fileSystemManager.getClusterId(); -936status.setStatus("Publishing Cluster ID " + clusterId + " in ZooKeeper"); -937 ZKClusterId.setClusterId(this.zooKeeper, fileSystemManager.getClusterId()); -938this.clusterId = clusterId.toString(); -939 -940// Precaution. Put in place the old hbck1 lock file to fence out old hbase1s running their -941// hbck1s against an hbase2 cluster; it could do damage. To skip this behavior, set -942// hbase.write.hbck1.lock.file to false. -943if (this.conf.getBoolean("hbase.write.hbck1.lock.file", true)) { -944 HBaseFsck.checkAndMarkRunningHbck(this.conf, -945 HBaseFsck.createLockRetryCounterFactory(this.conf).create()); -946} -947 -948status.setStatus("Initialize ServerManager and schedule SCP for crash servers"); -949// The below two managers must be created before loading procedures, as they will be used during -950// loading. -951this.serverManager = createServerManager(this); -952this.syncReplicationReplayWALManager = new SyncReplicationReplayWALManager(this); -953createProcedureExecutor(); -954@SuppressWarnings("rawtypes") -955MapClass? extends Procedure, ListProcedureMasterProcedureEnv procsByType = -956 procedureExecutor.getActiveProceduresNoCopy().stream() -957.collect(Collectors.groupingBy(p - p.getClass())); -958 -959 checkUnsupportedProcedure(procsByType); -960 -961// Create Assignment Manager -962this.assignmentManager = createAssignmentManager(this); -963this.assignmentManager.start(); -964// TODO: TRSP can perform as the sub procedure for other procedures, so even if it is marked as -965// completed, it could still be in the procedure list. This is a bit strange but is another -966// story, need to verify the implementation for ProcedureExecutor and ProcedureStore. -967 ListTransitRegionStateProcedure ritList = -968 procsByType.getOrDefault(TransitRegionStateProcedure.class, Collections.emptyList()).stream() -969.filter(p - !p.isFinished()).map(p - (TransitRegionStateProcedure) p) -970.collect(Collectors.toList()); -971 this.assignmentManager.setupRIT(ritList); -972 -973// Start RegionServerTracker with listing of servers found with exiting SCPs -- these should -974// be registered in the deadServers set -- and with the list of servernames out on the -975// filesystem that COULD BE 'alive' (we'll schedule SCPs for each and let