[7/7] hbase git commit: Amend HBASE-18026 ProtobufUtil seems to do extra array copying
Amend HBASE-18026 ProtobufUtil seems to do extra array copying If the ByteString is not a LiteralByteString, just do toByteArray(). Signed-off-by: Andrew PurtellProject: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/2cff94c4 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/2cff94c4 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/2cff94c4 Branch: refs/heads/branch-1.1 Commit: 2cff94c40933ce4f431f733b2de6d44e3e19c6f1 Parents: 26cb211 Author: Vincent Authored: Sun May 14 19:26:15 2017 -0700 Committer: Andrew Purtell Committed: Mon May 15 18:09:48 2017 -0700 -- .../main/java/com/google/protobuf/HBaseZeroCopyByteString.java | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/2cff94c4/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java -- diff --git a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java index 933a6e2..78ddd02 100644 --- a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java +++ b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java @@ -61,7 +61,7 @@ public final class HBaseZeroCopyByteString extends LiteralByteString { if (buf instanceof LiteralByteString) { return ((LiteralByteString) buf).bytes; } -throw new UnsupportedOperationException("Need a LiteralByteString, got a " -+ buf.getClass().getName()); +// In case it's BoundedByteString +return buf.toByteArray(); } }
[3/7] hbase git commit: Revert "HBASE-18026 ProtobufUtil seems to do extra array copying"
Revert "HBASE-18026 ProtobufUtil seems to do extra array copying" This reverts commit 58956316342b3eb90cb3d50ed74e4ad1914284f8. Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/37650775 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/37650775 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/37650775 Branch: refs/heads/master Commit: 37650775a59df5d3348ff6608807d5a8eb87f2a4 Parents: 6b60ba8 Author: Andrew PurtellAuthored: Mon May 15 18:06:33 2017 -0700 Committer: Andrew Purtell Committed: Mon May 15 18:06:33 2017 -0700 -- .../hadoop/hbase/protobuf/ProtobufUtil.java | 42 ++-- 1 file changed, 20 insertions(+), 22 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/37650775/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java index 5a6cd21..fcf2c34 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java @@ -17,8 +17,6 @@ */ package org.apache.hadoop.hbase.protobuf; -import static com.google.protobuf.HBaseZeroCopyByteString.zeroCopyGetBytes; - import java.io.IOException; import java.lang.reflect.Constructor; import java.lang.reflect.Method; @@ -369,7 +367,7 @@ public final class ProtobufUtil { */ public static Get toGet(final ClientProtos.Get proto) throws IOException { if (proto == null) return null; -byte[] row = zeroCopyGetBytes(proto.getRow()); +byte[] row = proto.getRow().toByteArray(); Get get = new Get(row); if (proto.hasCacheBlocks()) { get.setCacheBlocks(proto.getCacheBlocks()); @@ -466,7 +464,7 @@ public final class ProtobufUtil { MutationType type = proto.getMutateType(); assert type == MutationType.PUT: type.name(); long timestamp = proto.hasTimestamp()? proto.getTimestamp(): HConstants.LATEST_TIMESTAMP; -Put put = proto.hasRow() ? new Put(zeroCopyGetBytes(proto.getRow()), timestamp) : null; +Put put = proto.hasRow() ? new Put(proto.getRow().toByteArray(), timestamp) : null; int cellCount = proto.hasAssociatedCellCount()? proto.getAssociatedCellCount(): 0; if (cellCount > 0) { // The proto has metadata only and the data is separate to be found in the cellScanner. @@ -491,7 +489,7 @@ public final class ProtobufUtil { } // The proto has the metadata and the data itself for (ColumnValue column: proto.getColumnValueList()) { -byte[] family = zeroCopyGetBytes(column.getFamily()); +byte[] family = column.getFamily().toByteArray(); for (QualifierValue qv: column.getQualifierValueList()) { if (!qv.hasValue()) { throw new DoNotRetryIOException( @@ -510,7 +508,7 @@ public final class ProtobufUtil { allTagsBytes = qv.getTags().toByteArray(); if(qv.hasDeleteType()) { byte[] qual = qv.hasQualifier() ? qv.getQualifier().toByteArray() : null; - put.add(new KeyValue(zeroCopyGetBytes(proto.getRow()), family, qual, ts, + put.add(new KeyValue(proto.getRow().toByteArray(), family, qual, ts, fromDeleteType(qv.getDeleteType()), null, allTagsBytes)); } else { List tags = TagUtil.asList(allTagsBytes, 0, (short)allTagsBytes.length); @@ -519,8 +517,8 @@ public final class ProtobufUtil { } } else { if(qv.hasDeleteType()) { - byte[] qual = qv.hasQualifier() ? zeroCopyGetBytes(qv.getQualifier()) : null; - put.add(new KeyValue(zeroCopyGetBytes(proto.getRow()), family, qual, ts, + byte[] qual = qv.hasQualifier() ? qv.getQualifier().toByteArray() : null; + put.add(new KeyValue(proto.getRow().toByteArray(), family, qual, ts, fromDeleteType(qv.getDeleteType(; } else{ put.addImmutable(family, qualifier, ts, value); @@ -561,7 +559,7 @@ public final class ProtobufUtil { MutationType type = proto.getMutateType(); assert type == MutationType.DELETE : type.name(); long timestamp = proto.hasTimestamp() ? proto.getTimestamp() : HConstants.LATEST_TIMESTAMP; -Delete delete = proto.hasRow() ? new Delete(zeroCopyGetBytes(proto.getRow()), timestamp) : null; +Delete delete = proto.hasRow() ? new Delete(proto.getRow().toByteArray(), timestamp) : null; int cellCount = proto.hasAssociatedCellCount()?
[1/7] hbase git commit: HBASE-18043 Institute a hard limit for individual cell size that cannot be overridden by clients
Repository: hbase Updated Branches: refs/heads/branch-1 0a4528225 -> d8ef49506 refs/heads/branch-1.1 26cb211e1 -> 2cff94c40 refs/heads/branch-1.2 14ab4a9c4 -> 719ab32cc refs/heads/branch-1.3 36ebe05fc -> d18eb62d8 refs/heads/master 841bb0065 -> 37650775a HBASE-18043 Institute a hard limit for individual cell size that cannot be overridden by clients Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/29222669 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/29222669 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/29222669 Branch: refs/heads/branch-1 Commit: 292226690ac842be4513366633f31bba4fa62d34 Parents: 0a45282 Author: Andrew PurtellAuthored: Mon May 15 18:03:24 2017 -0700 Committer: Andrew Purtell Committed: Mon May 15 18:03:24 2017 -0700 -- .../src/main/resources/hbase-default.xml| 9 + .../hadoop/hbase/regionserver/HRegion.java | 11 +- .../hbase/regionserver/RSRpcServices.java | 35 -- .../hadoop/hbase/client/TestFromClientSide.java | 37 4 files changed, 89 insertions(+), 3 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/29222669/hbase-common/src/main/resources/hbase-default.xml -- diff --git a/hbase-common/src/main/resources/hbase-default.xml b/hbase-common/src/main/resources/hbase-default.xml index 2a74a6a..c571289 100644 --- a/hbase-common/src/main/resources/hbase-default.xml +++ b/hbase-common/src/main/resources/hbase-default.xml @@ -555,6 +555,15 @@ possible configurations would overwhelm and obscure the important. or less disables the check. +hbase.server.keyvalue.maxsize +10485760 +Maximum allowed size of an individual cell, inclusive of value and all key +components. A value of 0 or less disables the check. +The default value is 10MB. +This is a safety setting to protect the server from OOM situations. + + + hbase.client.scanner.timeout.period 6 Client scanner lease period in milliseconds. http://git-wip-us.apache.org/repos/asf/hbase/blob/29222669/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java index 8e1306f..1c44f6a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java @@ -218,6 +218,9 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi public static final String HREGION_UNASSIGN_FOR_FNFE = "hbase.hregion.unassign.for.fnfe"; public static final boolean DEFAULT_HREGION_UNASSIGN_FOR_FNFE = true; + public static final String HBASE_MAX_CELL_SIZE_KEY = "hbase.server.keyvalue.maxsize"; + public static final int DEFAULT_MAX_CELL_SIZE = 10485760; + /** * Longest time we'll wait on a sequenceid. * Sequenceid comes up out of the WAL subsystem. WAL subsystem can go bad or a test might use @@ -333,6 +336,10 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi // purge timeout, when a RPC call will be terminated by the RPC engine. final long maxBusyWaitDuration; + // Max cell size. If nonzero, the maximum allowed size for any given cell + // in bytes + final long maxCellSize; + // negative number indicates infinite timeout static final long DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L; final ExecutorService rowProcessorExecutor = Executors.newCachedThreadPool(); @@ -816,6 +823,8 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi conf.getBoolean(HConstants.ENABLE_CLIENT_BACKPRESSURE, HConstants.DEFAULT_ENABLE_CLIENT_BACKPRESSURE); +this.maxCellSize = conf.getLong(HBASE_MAX_CELL_SIZE_KEY, DEFAULT_MAX_CELL_SIZE); + boolean unassignForFNFE = conf.getBoolean(HREGION_UNASSIGN_FOR_FNFE, DEFAULT_HREGION_UNASSIGN_FOR_FNFE); if (unassignForFNFE) { @@ -8137,7 +8146,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi ClassSize.OBJECT + ClassSize.ARRAY + 46 * ClassSize.REFERENCE + 3 * Bytes.SIZEOF_INT + - (14 * Bytes.SIZEOF_LONG) + + (15 * Bytes.SIZEOF_LONG) + 5 * Bytes.SIZEOF_BOOLEAN); // woefully out of date - currently missing: http://git-wip-us.apache.org/repos/asf/hbase/blob/29222669/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
[5/7] hbase git commit: Amend HBASE-18026 ProtobufUtil seems to do extra array copying
Amend HBASE-18026 ProtobufUtil seems to do extra array copying If the ByteString is not a LiteralByteString, just do toByteArray(). Signed-off-by: Andrew PurtellProject: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/d18eb62d Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/d18eb62d Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/d18eb62d Branch: refs/heads/branch-1.3 Commit: d18eb62d8d0f02ea822090f409c2ded6f5279c3d Parents: 36ebe05 Author: Vincent Authored: Sun May 14 19:26:15 2017 -0700 Committer: Andrew Purtell Committed: Mon May 15 18:09:42 2017 -0700 -- .../main/java/com/google/protobuf/HBaseZeroCopyByteString.java | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/d18eb62d/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java -- diff --git a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java index 9d75612..d0acd4c 100644 --- a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java +++ b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java @@ -68,7 +68,7 @@ public final class HBaseZeroCopyByteString extends LiteralByteString { if (buf instanceof LiteralByteString) { return ((LiteralByteString) buf).bytes; } -throw new UnsupportedOperationException("Need a LiteralByteString, got a " -+ buf.getClass().getName()); +// In case it's BoundedByteString +return buf.toByteArray(); } }
[2/7] hbase git commit: HBASE-18043 Institute a hard limit for individual cell size that cannot be overridden by clients
HBASE-18043 Institute a hard limit for individual cell size that cannot be overridden by clients Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/6b60ba8a Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/6b60ba8a Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/6b60ba8a Branch: refs/heads/master Commit: 6b60ba8adebd531597d59d361b11fc1e32b40523 Parents: 841bb00 Author: Andrew PurtellAuthored: Mon May 15 18:03:33 2017 -0700 Committer: Andrew Purtell Committed: Mon May 15 18:03:33 2017 -0700 -- .../src/main/resources/hbase-default.xml| 9 + .../hadoop/hbase/regionserver/HRegion.java | 12 ++- .../hbase/regionserver/RSRpcServices.java | 35 -- .../hadoop/hbase/client/TestFromClientSide.java | 37 4 files changed, 90 insertions(+), 3 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/6b60ba8a/hbase-common/src/main/resources/hbase-default.xml -- diff --git a/hbase-common/src/main/resources/hbase-default.xml b/hbase-common/src/main/resources/hbase-default.xml index 40cf675..a6e37ef 100644 --- a/hbase-common/src/main/resources/hbase-default.xml +++ b/hbase-common/src/main/resources/hbase-default.xml @@ -532,6 +532,15 @@ possible configurations would overwhelm and obscure the important. or less disables the check. +hbase.server.keyvalue.maxsize +10485760 +Maximum allowed size of an individual cell, inclusive of value and all key +components. A value of 0 or less disables the check. +The default value is 10MB. +This is a safety setting to protect the server from OOM situations. + + + hbase.client.scanner.timeout.period 6 Client scanner lease period in milliseconds. http://git-wip-us.apache.org/repos/asf/hbase/blob/6b60ba8a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java index 4836dc8..91fb44b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java @@ -205,6 +205,9 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi public static final String HREGION_UNASSIGN_FOR_FNFE = "hbase.hregion.unassign.for.fnfe"; public static final boolean DEFAULT_HREGION_UNASSIGN_FOR_FNFE = true; + public static final String HBASE_MAX_CELL_SIZE_KEY = "hbase.server.keyvalue.maxsize"; + public static final int DEFAULT_MAX_CELL_SIZE = 10485760; + /** * This is the global default value for durability. All tables/mutations not * defining a durability or using USE_DEFAULT will default to this value. @@ -307,6 +310,10 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi // purge timeout, when a RPC call will be terminated by the RPC engine. final long maxBusyWaitDuration; + // Max cell size. If nonzero, the maximum allowed size for any given cell + // in bytes + final long maxCellSize; + // negative number indicates infinite timeout static final long DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L; final ExecutorService rowProcessorExecutor = Executors.newCachedThreadPool(); @@ -801,6 +808,9 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi false : conf.getBoolean(HConstants.ENABLE_CLIENT_BACKPRESSURE, HConstants.DEFAULT_ENABLE_CLIENT_BACKPRESSURE); + +this.maxCellSize = conf.getLong(HBASE_MAX_CELL_SIZE_KEY, DEFAULT_MAX_CELL_SIZE); + boolean unassignForFNFE = conf.getBoolean(HREGION_UNASSIGN_FOR_FNFE, DEFAULT_HREGION_UNASSIGN_FOR_FNFE); if (unassignForFNFE) { @@ -7613,7 +7623,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver, Regi ClassSize.OBJECT + ClassSize.ARRAY + 49 * ClassSize.REFERENCE + 2 * Bytes.SIZEOF_INT + - (14 * Bytes.SIZEOF_LONG) + + (15 * Bytes.SIZEOF_LONG) + 6 * Bytes.SIZEOF_BOOLEAN); // woefully out of date - currently missing: http://git-wip-us.apache.org/repos/asf/hbase/blob/6b60ba8a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
[6/7] hbase git commit: Amend HBASE-18026 ProtobufUtil seems to do extra array copying
Amend HBASE-18026 ProtobufUtil seems to do extra array copying If the ByteString is not a LiteralByteString, just do toByteArray(). Signed-off-by: Andrew PurtellProject: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/719ab32c Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/719ab32c Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/719ab32c Branch: refs/heads/branch-1.2 Commit: 719ab32cc6910cf393ce235315190ffd8dd7a3c8 Parents: 14ab4a9 Author: Vincent Authored: Sun May 14 19:26:15 2017 -0700 Committer: Andrew Purtell Committed: Mon May 15 18:09:46 2017 -0700 -- .../main/java/com/google/protobuf/HBaseZeroCopyByteString.java | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/719ab32c/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java -- diff --git a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java index 9d75612..d0acd4c 100644 --- a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java +++ b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java @@ -68,7 +68,7 @@ public final class HBaseZeroCopyByteString extends LiteralByteString { if (buf instanceof LiteralByteString) { return ((LiteralByteString) buf).bytes; } -throw new UnsupportedOperationException("Need a LiteralByteString, got a " -+ buf.getClass().getName()); +// In case it's BoundedByteString +return buf.toByteArray(); } }
[4/7] hbase git commit: Amend HBASE-18026 ProtobufUtil seems to do extra array copying
Amend HBASE-18026 ProtobufUtil seems to do extra array copying If the ByteString is not a LiteralByteString, just do toByteArray(). Signed-off-by: Andrew PurtellProject: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/d8ef4950 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/d8ef4950 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/d8ef4950 Branch: refs/heads/branch-1 Commit: d8ef495063ebf794164603d0739fcb27dce106fa Parents: 2922266 Author: Vincent Authored: Sun May 14 19:26:15 2017 -0700 Committer: Andrew Purtell Committed: Mon May 15 18:07:53 2017 -0700 -- .../main/java/com/google/protobuf/HBaseZeroCopyByteString.java | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/d8ef4950/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java -- diff --git a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java index 9d75612..d0acd4c 100644 --- a/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java +++ b/hbase-protocol/src/main/java/com/google/protobuf/HBaseZeroCopyByteString.java @@ -68,7 +68,7 @@ public final class HBaseZeroCopyByteString extends LiteralByteString { if (buf instanceof LiteralByteString) { return ((LiteralByteString) buf).bytes; } -throw new UnsupportedOperationException("Need a LiteralByteString, got a " -+ buf.getClass().getName()); +// In case it's BoundedByteString +return buf.toByteArray(); } }
hbase git commit: HBASE-18044 Fix bug in report-flakies.py where if counter is set outside for loop, it cannot be overwritten inside loop.
Repository: hbase Updated Branches: refs/heads/master 341223d86 -> 841bb0065 HBASE-18044 Fix bug in report-flakies.py where if counter is set outside for loop, it cannot be overwritten inside loop. http://stackoverflow.com/questions/7537439/how-to-increment-a-variable-on-a-for-loop-in-jinja-template Change-Id: Ic404d6360fb646e6490d1487374520f9550b76b4 Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/841bb006 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/841bb006 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/841bb006 Branch: refs/heads/master Commit: 841bb00655b3eb58e7d76d945e24cfeadfdb11de Parents: 341223d Author: Apekshit SharmaAuthored: Fri May 12 02:25:19 2017 -0700 Committer: Apekshit Sharma Committed: Mon May 15 14:15:57 2017 -0700 -- dev-support/flaky-dashboard-template.html | 11 --- 1 file changed, 4 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/841bb006/dev-support/flaky-dashboard-template.html -- diff --git a/dev-support/flaky-dashboard-template.html b/dev-support/flaky-dashboard-template.html index fff7c82..b980b5d 100644 --- a/dev-support/flaky-dashboard-template.html +++ b/dev-support/flaky-dashboard-template.html @@ -55,22 +55,19 @@ List of Jobs -{% set counter = 0 %} {% for url in results %} -{% set counter = counter + 1 %} -{{ url |e }} +{{ url |e }} {% endfor %} Results -{% set counter = 0 %} {% for url in results %} {% set result = results[url] %} +{% set url_counter = loop.index %} {# Dedup ids since test names may duplicate across urls #} -{% set counter = counter + 1 %} - + {{ url |e }} Go to https://jenkins.io/sites/default/files/jenkins_favicon.ico;> @@ -108,7 +105,7 @@ {{ failed|length }} / {{ timeout|length }} / {{ hanging|length }} -{% set id = "details_" ~ test ~ "_" ~ counter %} +{% set id = "details_" ~ test ~ "_" ~ url_counter %} show/hide
hbase git commit: Two SPLIT requests came in on top of each other; the second failed because it saw parent region was SPLIT. I 'fixed' this before but my fix was in the wrong place
Repository: hbase Updated Branches: refs/heads/HBASE-14614 50511ad24 -> 9460eaf1b Two SPLIT requests came in on top of each other; the second failed because it saw parent region was SPLIT. I 'fixed' this before but my fix was in the wrong place Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/9460eaf1 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/9460eaf1 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/9460eaf1 Branch: refs/heads/HBASE-14614 Commit: 9460eaf1bfd04b74bc7e4da7405c1601bd02b791 Parents: 50511ad2 Author: Michael StackAuthored: Mon May 15 09:49:30 2017 -0700 Committer: Michael Stack Committed: Mon May 15 09:49:30 2017 -0700 -- .../hbase/procedure2/ProcedureExecutor.java | 2 +- .../assignment/SplitTableRegionProcedure.java | 20 .../hadoop/hbase/regionserver/HRegion.java | 5 + 3 files changed, 18 insertions(+), 9 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/9460eaf1/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java -- diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java index 258e268..bc73453 100644 --- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java +++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java @@ -1147,7 +1147,7 @@ public class ProcedureExecutor { scheduler.yield(proc); break; case LOCK_EVENT_WAIT: - // someone will wake us up when the lock is available + // Someone will wake us up when the lock is available LOG.debug(lockState + " " + proc); break; default: http://git-wip-us.apache.org/repos/asf/hbase/blob/9460eaf1/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.java index 6815e9f..7ebe769 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.java @@ -341,6 +341,15 @@ public class SplitTableRegionProcedure if (node != null) { parentHRI = node.getRegionInfo(); + // Lookup the parent HRI state from the AM, which has the latest updated info. + // Protect against the case where concurrent SPLIT requests came in. Check a SPLIT + // did not just run. + if (parentHRI.isSplit() || parentHRI.isOffline()) { +setFailure(new IOException("Split " + parentHRI.getRegionNameAsString() + " FAILED because " + +"offline/split already.")); +return false; + } + // expected parent to be online or closed if (!node.isInState(EXPECTED_SPLIT_STATES)) { // We may have SPLIT already? @@ -350,13 +359,6 @@ public class SplitTableRegionProcedure return false; } - // lookup the parent HRI state from the AM, which has the latest updated info. - if (parentHRI.isSplit() || parentHRI.isOffline()) { -setFailure(new IOException("Split " + parentHRI.getRegionNameAsString() + " FAILED because " + -"offline/split already.")); -return false; - } - // Ask the remote regionserver if this region is splittable. If we get an IOE, report it // along w/ the failure so can see why we are not splittable at this time. IOException splittableCheckIOE = null; @@ -365,7 +367,9 @@ public class SplitTableRegionProcedure GetRegionInfoResponse response = Util.getRegionInfoResponse(env, node.getRegionLocation(), node.getRegionInfo()); splittable = response.hasSplittable() && response.getSplittable(); -LOG.info("REMOVE splittable " + splittable + " " + this + " " + node); +if (LOG.isDebugEnabled()) { + LOG.debug("Splittable=" + splittable + " " + this + " " + node.toShortString()); +} } catch (IOException e) { splittableCheckIOE = e; } http://git-wip-us.apache.org/repos/asf/hbase/blob/9460eaf1/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java -- diff --git
[07/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html index ca92b5e..c703b08 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html @@ -69,15 +69,15 @@ 061 requiredArguments = { 062@org.jamon.annotations.Argument(name = "master", type = "HMaster")}, 063 optionalArguments = { -064@org.jamon.annotations.Argument(name = "serverManager", type = "ServerManager"), -065@org.jamon.annotations.Argument(name = "format", type = "String"), +064@org.jamon.annotations.Argument(name = "deadServers", type = "SetServerName"), +065@org.jamon.annotations.Argument(name = "frags", type = "MapString,Integer"), 066@org.jamon.annotations.Argument(name = "catalogJanitorEnabled", type = "boolean"), -067@org.jamon.annotations.Argument(name = "servers", type = "ListServerName"), -068@org.jamon.annotations.Argument(name = "filter", type = "String"), -069@org.jamon.annotations.Argument(name = "assignmentManager", type = "AssignmentManager"), -070@org.jamon.annotations.Argument(name = "deadServers", type = "SetServerName"), -071@org.jamon.annotations.Argument(name = "frags", type = "MapString,Integer"), -072@org.jamon.annotations.Argument(name = "metaLocation", type = "ServerName")}) +067@org.jamon.annotations.Argument(name = "format", type = "String"), +068@org.jamon.annotations.Argument(name = "assignmentManager", type = "AssignmentManager"), +069@org.jamon.annotations.Argument(name = "metaLocation", type = "ServerName"), +070@org.jamon.annotations.Argument(name = "filter", type = "String"), +071@org.jamon.annotations.Argument(name = "servers", type = "ListServerName"), +072@org.jamon.annotations.Argument(name = "serverManager", type = "ServerManager")}) 073public class MasterStatusTmpl 074 extends org.jamon.AbstractTemplateProxy 075{ @@ -118,40 +118,40 @@ 110 return m_master; 111} 112private HMaster m_master; -113// 28, 1 -114public void setServerManager(ServerManager serverManager) +113// 24, 1 +114public void setDeadServers(SetServerName deadServers) 115{ -116 // 28, 1 -117 m_serverManager = serverManager; -118 m_serverManager__IsNotDefault = true; +116 // 24, 1 +117 m_deadServers = deadServers; +118 m_deadServers__IsNotDefault = true; 119} -120public ServerManager getServerManager() +120public SetServerName getDeadServers() 121{ -122 return m_serverManager; +122 return m_deadServers; 123} -124private ServerManager m_serverManager; -125public boolean getServerManager__IsNotDefault() +124private SetServerName m_deadServers; +125public boolean getDeadServers__IsNotDefault() 126{ -127 return m_serverManager__IsNotDefault; +127 return m_deadServers__IsNotDefault; 128} -129private boolean m_serverManager__IsNotDefault; -130// 27, 1 -131public void setFormat(String format) +129private boolean m_deadServers__IsNotDefault; +130// 21, 1 +131public void setFrags(MapString,Integer frags) 132{ -133 // 27, 1 -134 m_format = format; -135 m_format__IsNotDefault = true; +133 // 21, 1 +134 m_frags = frags; +135 m_frags__IsNotDefault = true; 136} -137public String getFormat() +137public MapString,Integer getFrags() 138{ -139 return m_format; +139 return m_frags; 140} -141private String m_format; -142public boolean getFormat__IsNotDefault() +141private MapString,Integer m_frags; +142public boolean getFrags__IsNotDefault() 143{ -144 return m_format__IsNotDefault; +144 return m_frags__IsNotDefault; 145} -146private boolean m_format__IsNotDefault; +146private boolean m_frags__IsNotDefault; 147// 25, 1 148public void setCatalogJanitorEnabled(boolean catalogJanitorEnabled) 149{ @@ -169,108 +169,108 @@ 161 return m_catalogJanitorEnabled__IsNotDefault; 162} 163private boolean m_catalogJanitorEnabled__IsNotDefault; -164// 23, 1 -165public void setServers(ListServerName servers) +164// 27, 1 +165public void setFormat(String format) 166{ -167 // 23, 1 -168 m_servers = servers; -169 m_servers__IsNotDefault = true; +167 // 27, 1 +168 m_format = format; +169 m_format__IsNotDefault = true; 170} -171public ListServerName getServers() +171public String getFormat() 172{ -173
[38/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.html b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.html new file mode 100644 index 000..977396d --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.html @@ -0,0 +1,536 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +SimpleRpcServerResponder (Apache HBase 2.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.ipc +Class SimpleRpcServerResponder + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true; title="class or interface in java.lang">java.lang.Thread + + +org.apache.hadoop.hbase.ipc.SimpleRpcServerResponder + + + + + + + + + +All Implemented Interfaces: +http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true; title="class or interface in java.lang">Runnable + + + +@InterfaceAudience.Private +class SimpleRpcServerResponder +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true; title="class or interface in java.lang">Thread +Sends responses of RPC back to clients. + + + + + + + + + + + +Nested Class Summary + + + + +Nested classes/interfaces inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true; title="class or interface in java.lang">Thread +http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.State.html?is-external=true; title="class or interface in java.lang">Thread.State, http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.UncaughtExceptionHandler.html?is-external=true; title="class or interface in java.lang">Thread.UncaughtExceptionHandler + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +private SimpleRpcServer +simpleRpcServer + + +private http://docs.oracle.com/javase/8/docs/api/java/nio/channels/Selector.html?is-external=true; title="class or interface in java.nio.channels">Selector +writeSelector + + +private http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetSimpleServerRpcConnection +writingCons + + + + + + +Fields inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true; title="class or interface in java.lang">Thread +http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true#MAX_PRIORITY; title="class or interface in java.lang">MAX_PRIORITY, http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true#MIN_PRIORITY; title="class or interface in java.lang">MIN_PRIORITY, http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true#NORM_PRIORITY; title="class or interface in java.lang">NORM_PRIORITY + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +SimpleRpcServerResponder(SimpleRpcServersimpleRpcServer) + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +private void +doAsyncWrite(http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SelectionKey.html?is-external=true; title="class or interface in java.nio.channels">SelectionKeykey) + + +(package private) void +doRespond(SimpleServerCallcall) + + +private void +doRunLoop() + + +private boolean +processAllResponses(SimpleServerRpcConnectionconnection) +Process all the responses for this connection + + + +(package private) boolean +processResponse(SimpleServerCallcall) +Process the response
[46/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html b/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html deleted file mode 100644 index b4590eb..000 --- a/devapidocs/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html +++ /dev/null @@ -1,507 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - - - - -NettyRpcServer.NettyConnection (Apache HBase 2.0.0-SNAPSHOT API) - - - - - -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; -var altColor = "altColor"; -var rowColor = "rowColor"; -var tableTab = "tableTab"; -var activeTableTab = "activeTableTab"; - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -PrevClass -NextClass - - -Frames -NoFrames - - -AllClasses - - - - - - - -Summary: -Nested| -Field| -Constr| -Method - - -Detail: -Field| -Constr| -Method - - - - - - - - -org.apache.hadoop.hbase.ipc -Class NettyRpcServer.NettyConnection - - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object - - -org.apache.hadoop.hbase.ipc.RpcServer.Connection - - -org.apache.hadoop.hbase.ipc.NettyRpcServer.NettyConnection - - - - - - - - - -All Implemented Interfaces: -http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable, http://docs.oracle.com/javase/8/docs/api/java/lang/AutoCloseable.html?is-external=true; title="class or interface in java.lang">AutoCloseable - - -Enclosing class: -NettyRpcServer - - - -public class NettyRpcServer.NettyConnection -extends RpcServer.Connection - - - - - - - - - - - -Field Summary - -Fields - -Modifier and Type -Field and Description - - -protected io.netty.channel.Channel -channel - - - - - - -Fields inherited from classorg.apache.hadoop.hbase.ipc.RpcServer.Connection -addr, attemptingUser, authenticatedWithFallback, authFailedCall, authFailedResponse, authMethod, AUTHORIZATION_FAILED_CALLID, callCleanup, codec, compressionCodec, CONNECTION_HEADER_RESPONSE_CALLID, connectionHeader, connectionHeaderRead, connectionPreambleRead, cryptoAES, hostAddress, remotePort, retryImmediatelySupported, SASL_CALLID, saslCall, saslContextEstablished, saslServer, service, setConnectionHeaderResponseCall, skipInitialSaslHandshake, ugi, useCryptoAesWrap, user, useSasl, useWrap - - - - - - - - -Constructor Summary - -Constructors - -Constructor and Description - - -NettyConnection(io.netty.channel.Channelchannel) - - - - - - - - - -Method Summary - -All MethodsInstance MethodsConcrete Methods - -Modifier and Type -Method and Description - - -void -close() - - -ServerCall -createCall(intid, - org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingServiceservice, - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptormd, - org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeaderheader, - org.apache.hadoop.hbase.shaded.com.google.protobuf.Messageparam, - CellScannercellScanner, - RpcServer.Connectionconnection, - longsize, - org.apache.htrace.TraceInfotinfo, - http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddressremoteAddress, - inttimeout, - RpcServer.CallCleanupreqCleanup) - - -private void -doBadPreambleHandling(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringmsg) - - -private void -doBadPreambleHandling(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringmsg, - http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exceptione) - - -boolean -isConnectionOpen() - - -(package private) void -process(io.netty.buffer.ByteBufbuf) - - -(package private) void -process(ByteBuffbuf) - - -(package
[26/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html deleted file mode 100644 index 109b5f3..000 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.NettyConnection.html +++ /dev/null @@ -1,567 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - -Source code - - - - -001/** -002 * Licensed to the Apache Software Foundation (ASF) under one -003 * or more contributor license agreements. See the NOTICE file -004 * distributed with this work for additional information -005 * regarding copyright ownership. The ASF licenses this file -006 * to you under the Apache License, Version 2.0 (the -007 * "License"); you may not use this file except in compliance -008 * with the License. You may obtain a copy of the License at -009 * -010 * http://www.apache.org/licenses/LICENSE-2.0 -011 * -012 * Unless required by applicable law or agreed to in writing, software -013 * distributed under the License is distributed on an "AS IS" BASIS, -014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -015 * See the License for the specific language governing permissions and -016 * limitations under the License. -017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import io.netty.bootstrap.ServerBootstrap; -022import io.netty.buffer.ByteBuf; -023import io.netty.buffer.PooledByteBufAllocator; -024import io.netty.buffer.Unpooled; -025import io.netty.channel.Channel; -026import io.netty.channel.ChannelFuture; -027import io.netty.channel.ChannelFutureListener; -028import io.netty.channel.ChannelHandlerContext; -029import io.netty.channel.ChannelInboundHandlerAdapter; -030import io.netty.channel.ChannelInitializer; -031import io.netty.channel.ChannelOption; -032import io.netty.channel.ChannelOutboundHandlerAdapter; -033import io.netty.channel.ChannelPipeline; -034import io.netty.channel.ChannelPromise; -035import io.netty.channel.EventLoopGroup; -036import io.netty.channel.epoll.EpollEventLoopGroup; -037import io.netty.channel.epoll.EpollServerSocketChannel; -038import io.netty.channel.group.ChannelGroup; -039import io.netty.channel.group.DefaultChannelGroup; -040import io.netty.channel.nio.NioEventLoopGroup; -041import io.netty.channel.socket.SocketChannel; -042import io.netty.channel.socket.nio.NioServerSocketChannel; -043import io.netty.handler.codec.ByteToMessageDecoder; -044import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -045import io.netty.util.concurrent.GlobalEventExecutor; -046 -047import java.io.IOException; -048import java.io.InterruptedIOException; -049import java.net.InetAddress; -050import java.net.InetSocketAddress; -051import java.nio.ByteBuffer; -052import java.util.Arrays; -053import java.util.List; -054import java.util.concurrent.CountDownLatch; -055 -056import org.apache.commons.logging.Log; -057import org.apache.commons.logging.LogFactory; -058import org.apache.hadoop.conf.Configuration; -059import org.apache.hadoop.hbase.CellScanner; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -063import org.apache.hadoop.hbase.nio.ByteBuff; -064import org.apache.hadoop.hbase.nio.SingleByteBuff; -065import org.apache.hadoop.hbase.security.AccessDeniedException; -066import org.apache.hadoop.hbase.security.AuthMethod; -067import org.apache.hadoop.hbase.security.HBasePolicyProvider; -068import org.apache.hadoop.hbase.security.SaslStatus; -069import org.apache.hadoop.hbase.security.SaslUtil; -070import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -073import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -074import org.apache.hadoop.hbase.util.Bytes; -075import org.apache.hadoop.hbase.util.JVM; -076import org.apache.hadoop.hbase.util.Pair; -077import org.apache.hadoop.io.IntWritable; -078import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -079import org.apache.htrace.TraceInfo; -080 -081/** -082 * An RPC server with Netty4 implementation. -083 * -084 */ -085public class NettyRpcServer extends RpcServer { -086 -087 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -088 -089 protected final InetSocketAddress bindAddress; -090 -091 private final CountDownLatch closed = new CountDownLatch(1); -092 private final Channel serverChannel; -093 private final ChannelGroup allChannels = new
[10/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html index 7f61b54..35e8890 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.html @@ -23,1363 +23,690 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import java.io.IOException; -022import java.io.InputStream; -023import java.net.BindException; -024import java.net.InetAddress; -025import java.net.InetSocketAddress; -026import java.net.ServerSocket; -027import java.net.Socket; -028import java.net.SocketException; -029import java.net.UnknownHostException; -030import java.nio.ByteBuffer; -031import java.nio.channels.CancelledKeyException; -032import java.nio.channels.ClosedChannelException; -033import java.nio.channels.GatheringByteChannel; -034import java.nio.channels.ReadableByteChannel; -035import java.nio.channels.SelectionKey; -036import java.nio.channels.Selector; -037import java.nio.channels.ServerSocketChannel; -038import java.nio.channels.SocketChannel; -039import java.util.ArrayList; -040import java.util.Arrays; -041import java.util.Collections; -042import java.util.Iterator; -043import java.util.List; -044import java.util.Set; -045import java.util.Timer; -046import java.util.TimerTask; -047import java.util.concurrent.ConcurrentHashMap; -048import java.util.concurrent.ConcurrentLinkedDeque; -049import java.util.concurrent.ExecutorService; -050import java.util.concurrent.Executors; -051import java.util.concurrent.LinkedBlockingQueue; -052import java.util.concurrent.atomic.AtomicInteger; -053import java.util.concurrent.atomic.LongAdder; -054import java.util.concurrent.locks.Lock; -055import java.util.concurrent.locks.ReentrantLock; -056 -057import org.apache.hadoop.conf.Configuration; -058import org.apache.hadoop.hbase.CellScanner; -059import org.apache.hadoop.hbase.DoNotRetryIOException; -060import org.apache.hadoop.hbase.HBaseIOException; -061import org.apache.hadoop.hbase.HConstants; -062import org.apache.hadoop.hbase.Server; -063import org.apache.hadoop.hbase.classification.InterfaceAudience; -064import org.apache.hadoop.hbase.classification.InterfaceStability; -065import org.apache.hadoop.hbase.client.VersionInfoUtil; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -068import org.apache.hadoop.hbase.nio.ByteBuff; -069import org.apache.hadoop.hbase.nio.SingleByteBuff; -070import org.apache.hadoop.hbase.security.AccessDeniedException; -071import org.apache.hadoop.hbase.security.AuthMethod; -072import org.apache.hadoop.hbase.security.HBasePolicyProvider; -073import org.apache.hadoop.hbase.security.SaslStatus; -074import org.apache.hadoop.hbase.security.SaslUtil; -075import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -076import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -077import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -078import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -079import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -081import org.apache.hadoop.hbase.util.Bytes; -082import org.apache.hadoop.hbase.util.Pair; -083import org.apache.hadoop.hbase.util.Threads; -084import org.apache.hadoop.io.IOUtils; -085import org.apache.hadoop.io.IntWritable; -086import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -087import org.apache.hadoop.util.StringUtils; -088import org.apache.htrace.TraceInfo; -089 -090import com.google.common.util.concurrent.ThreadFactoryBuilder; +018package org.apache.hadoop.hbase.ipc; +019 +020import java.io.IOException; +021import java.net.BindException; +022import java.net.InetSocketAddress; +023import java.net.ServerSocket; +024import java.net.SocketException; +025import java.net.UnknownHostException; +026import java.nio.channels.CancelledKeyException; +027import java.nio.channels.GatheringByteChannel; +028import java.nio.channels.SelectionKey; +029import java.nio.channels.Selector; +030import java.nio.channels.ServerSocketChannel; +031import java.nio.channels.SocketChannel; +032import java.util.Collections; +033import java.util.Iterator; +034import java.util.List; +035import java.util.Set; +036import java.util.Timer; +037import java.util.TimerTask; +038import java.util.concurrent.ConcurrentHashMap; +039import java.util.concurrent.ExecutorService;
[12/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html index 7f61b54..35e8890 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html @@ -23,1363 +23,690 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import java.io.IOException; -022import java.io.InputStream; -023import java.net.BindException; -024import java.net.InetAddress; -025import java.net.InetSocketAddress; -026import java.net.ServerSocket; -027import java.net.Socket; -028import java.net.SocketException; -029import java.net.UnknownHostException; -030import java.nio.ByteBuffer; -031import java.nio.channels.CancelledKeyException; -032import java.nio.channels.ClosedChannelException; -033import java.nio.channels.GatheringByteChannel; -034import java.nio.channels.ReadableByteChannel; -035import java.nio.channels.SelectionKey; -036import java.nio.channels.Selector; -037import java.nio.channels.ServerSocketChannel; -038import java.nio.channels.SocketChannel; -039import java.util.ArrayList; -040import java.util.Arrays; -041import java.util.Collections; -042import java.util.Iterator; -043import java.util.List; -044import java.util.Set; -045import java.util.Timer; -046import java.util.TimerTask; -047import java.util.concurrent.ConcurrentHashMap; -048import java.util.concurrent.ConcurrentLinkedDeque; -049import java.util.concurrent.ExecutorService; -050import java.util.concurrent.Executors; -051import java.util.concurrent.LinkedBlockingQueue; -052import java.util.concurrent.atomic.AtomicInteger; -053import java.util.concurrent.atomic.LongAdder; -054import java.util.concurrent.locks.Lock; -055import java.util.concurrent.locks.ReentrantLock; -056 -057import org.apache.hadoop.conf.Configuration; -058import org.apache.hadoop.hbase.CellScanner; -059import org.apache.hadoop.hbase.DoNotRetryIOException; -060import org.apache.hadoop.hbase.HBaseIOException; -061import org.apache.hadoop.hbase.HConstants; -062import org.apache.hadoop.hbase.Server; -063import org.apache.hadoop.hbase.classification.InterfaceAudience; -064import org.apache.hadoop.hbase.classification.InterfaceStability; -065import org.apache.hadoop.hbase.client.VersionInfoUtil; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -068import org.apache.hadoop.hbase.nio.ByteBuff; -069import org.apache.hadoop.hbase.nio.SingleByteBuff; -070import org.apache.hadoop.hbase.security.AccessDeniedException; -071import org.apache.hadoop.hbase.security.AuthMethod; -072import org.apache.hadoop.hbase.security.HBasePolicyProvider; -073import org.apache.hadoop.hbase.security.SaslStatus; -074import org.apache.hadoop.hbase.security.SaslUtil; -075import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -076import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -077import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -078import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -079import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -081import org.apache.hadoop.hbase.util.Bytes; -082import org.apache.hadoop.hbase.util.Pair; -083import org.apache.hadoop.hbase.util.Threads; -084import org.apache.hadoop.io.IOUtils; -085import org.apache.hadoop.io.IntWritable; -086import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -087import org.apache.hadoop.util.StringUtils; -088import org.apache.htrace.TraceInfo; -089 -090import com.google.common.util.concurrent.ThreadFactoryBuilder; +018package org.apache.hadoop.hbase.ipc; +019 +020import java.io.IOException; +021import java.net.BindException; +022import java.net.InetSocketAddress; +023import java.net.ServerSocket; +024import java.net.SocketException; +025import java.net.UnknownHostException; +026import java.nio.channels.CancelledKeyException; +027import java.nio.channels.GatheringByteChannel; +028import java.nio.channels.SelectionKey; +029import java.nio.channels.Selector; +030import java.nio.channels.ServerSocketChannel; +031import java.nio.channels.SocketChannel; +032import java.util.Collections; +033import java.util.Iterator; +034import java.util.List; +035import java.util.Set; +036import java.util.Timer; +037import java.util.TimerTask; +038import java.util.concurrent.ConcurrentHashMap;
[34/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/package-summary.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/package-summary.html b/devapidocs/org/apache/hadoop/hbase/ipc/package-summary.html index 2290619..33faadd 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/package-summary.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/package-summary.html @@ -388,13 +388,19 @@ An RPC server with Netty4 implementation. - + NettyServerCall Datastructure that holds all necessary to a method invocation and then afterward, carries the result. + +NettyServerRpcConnection + +RpcConnection implementation for netty rpc server. + + RpcClientFactory @@ -465,10 +471,6 @@ -RpcServer.ByteBuffByteInput - - - RpcServerFactory @@ -479,13 +481,23 @@ -ServerCall +ServerCallT extends ServerRpcConnection Datastructure that holds all necessary to a method invocation and then afterward, carries the result. +ServerRpcConnection + +Reads calls from a connection and queues them for handling. + + + +ServerRpcConnection.ByteBuffByteInput + + + ServerRpcController Used for server-side protobuf RPC service invocations. @@ -505,12 +517,24 @@ +SimpleRpcServerResponder + +Sends responses of RPC back to clients. + + + SimpleServerCall Datastructure that holds all necessary to a method invocation and then afterward, carries the result. + +SimpleServerRpcConnection + +Reads calls from a connection and queues them for handling. + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/package-tree.html b/devapidocs/org/apache/hadoop/hbase/ipc/package-tree.html index 2246697..a83f869 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/package-tree.html @@ -108,7 +108,7 @@ org.apache.hadoop.hbase.ipc.BufferChain org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput -org.apache.hadoop.hbase.ipc.RpcServer.ByteBuffByteInput +org.apache.hadoop.hbase.ipc.ServerRpcConnection.ByteBuffByteInput org.apache.hadoop.hbase.ipc.Call @@ -206,19 +206,19 @@ org.apache.hadoop.hbase.ipc.RpcServer.BlockingServiceAndInterface -org.apache.hadoop.hbase.ipc.RpcServer.Connection (implements java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable) - -org.apache.hadoop.hbase.ipc.NettyRpcServer.NettyConnection -org.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection - - org.apache.hadoop.hbase.ipc.RpcServerFactory -org.apache.hadoop.hbase.ipc.ServerCall (implements org.apache.hadoop.hbase.ipc.RpcCall) +org.apache.hadoop.hbase.ipc.ServerCallT (implements org.apache.hadoop.hbase.ipc.RpcCall) org.apache.hadoop.hbase.ipc.NettyServerCall org.apache.hadoop.hbase.ipc.SimpleServerCall +org.apache.hadoop.hbase.ipc.ServerRpcConnection (implements java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable) + +org.apache.hadoop.hbase.ipc.NettyServerRpcConnection +org.apache.hadoop.hbase.ipc.SimpleServerRpcConnection + + org.apache.hadoop.hbase.ipc.ServerRpcController (implements com.google.protobuf.RpcController) org.apache.hadoop.hbase.ipc.SimpleRpcServer.ConnectionManager org.apache.hadoop.hbase.ipc.SimpleRpcServer.Listener.Reader (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true; title="class or interface in java.lang">Runnable) @@ -231,7 +231,7 @@ org.apache.hadoop.hbase.ipc.SimpleRpcServer.Listener -org.apache.hadoop.hbase.ipc.SimpleRpcServer.Responder +org.apache.hadoop.hbase.ipc.SimpleRpcServerResponder java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable (implements java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) @@ -347,8 +347,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) org.apache.hadoop.hbase.ipc.CallEvent.Type -org.apache.hadoop.hbase.ipc.MetricsHBaseServerSourceFactoryImpl.SourceStorage org.apache.hadoop.hbase.ipc.BufferCallBeforeInitHandler.BufferCallAction
[21/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html index 9e1c66c..27ccbd5 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.CallCleanup.html @@ -28,1601 +28,787 @@ 020 021import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHORIZATION; 022 -023import java.io.ByteArrayInputStream; -024import java.io.ByteArrayOutputStream; -025import java.io.Closeable; -026import java.io.DataOutputStream; -027import java.io.IOException; -028import java.net.InetAddress; -029import java.net.InetSocketAddress; -030import java.nio.ByteBuffer; -031import java.nio.channels.Channels; -032import java.nio.channels.GatheringByteChannel; -033import java.nio.channels.ReadableByteChannel; -034import java.nio.channels.WritableByteChannel; -035import java.security.GeneralSecurityException; -036import java.security.PrivilegedExceptionAction; -037import java.util.ArrayList; -038import java.util.HashMap; -039import java.util.List; -040import java.util.Map; -041import java.util.Properties; -042import java.util.concurrent.atomic.LongAdder; -043 -044import javax.security.sasl.Sasl; -045import javax.security.sasl.SaslException; -046import javax.security.sasl.SaslServer; -047 -048import org.apache.commons.crypto.cipher.CryptoCipherFactory; -049import org.apache.commons.crypto.random.CryptoRandom; -050import org.apache.commons.crypto.random.CryptoRandomFactory; -051import org.apache.commons.logging.Log; -052import org.apache.commons.logging.LogFactory; -053import org.apache.hadoop.conf.Configuration; -054import org.apache.hadoop.hbase.CallQueueTooBigException; -055import org.apache.hadoop.hbase.CellScanner; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseInterfaceAudience; -058import org.apache.hadoop.hbase.HConstants; -059import org.apache.hadoop.hbase.Server; -060import org.apache.hadoop.hbase.classification.InterfaceAudience; -061import org.apache.hadoop.hbase.classification.InterfaceStability; -062import org.apache.hadoop.hbase.client.VersionInfoUtil; -063import org.apache.hadoop.hbase.codec.Codec; -064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -066import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -067import org.apache.hadoop.hbase.io.ByteBufferPool; -068import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -069import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -070import org.apache.hadoop.hbase.monitoring.TaskMonitor; -071import org.apache.hadoop.hbase.nio.ByteBuff; -072import org.apache.hadoop.hbase.nio.MultiByteBuff; -073import org.apache.hadoop.hbase.nio.SingleByteBuff; -074import org.apache.hadoop.hbase.regionserver.RSRpcServices; -075import org.apache.hadoop.hbase.security.AccessDeniedException; -076import org.apache.hadoop.hbase.security.AuthMethod; -077import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -078import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -080import org.apache.hadoop.hbase.security.SaslStatus; -081import org.apache.hadoop.hbase.security.SaslUtil; -082import org.apache.hadoop.hbase.security.User; -083import org.apache.hadoop.hbase.security.UserProvider; -084import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -085import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -086import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; -087import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; -088import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -089import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -090import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -091import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; -092import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; -093import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; -094import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; -099import
[35/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/class-use/ServerRpcConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/ServerRpcConnection.html b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/ServerRpcConnection.html new file mode 100644 index 000..4bf285a --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/ServerRpcConnection.html @@ -0,0 +1,204 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +Uses of Class org.apache.hadoop.hbase.ipc.ServerRpcConnection (Apache HBase 2.0.0-SNAPSHOT API) + + + + + + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + + +Uses of Classorg.apache.hadoop.hbase.ipc.ServerRpcConnection + + + + + +Packages that use ServerRpcConnection + +Package +Description + + + +org.apache.hadoop.hbase.ipc + +Tools to help define network clients and servers. + + + + + + + + + + +Uses of ServerRpcConnection in org.apache.hadoop.hbase.ipc + +Classes in org.apache.hadoop.hbase.ipc with type parameters of type ServerRpcConnection + +Modifier and Type +Class and Description + + + +(package private) class +ServerCallT extends ServerRpcConnection +Datastructure that holds all necessary to a method invocation and then afterward, carries + the result. + + + + + +Subclasses of ServerRpcConnection in org.apache.hadoop.hbase.ipc + +Modifier and Type +Class and Description + + + +(package private) class +NettyServerRpcConnection +RpcConnection implementation for netty rpc server. + + + +(package private) class +SimpleServerRpcConnection +Reads calls from a connection and queues them for handling. + + + + + +Fields in org.apache.hadoop.hbase.ipc declared as ServerRpcConnection + +Modifier and Type +Field and Description + + + +protected T +ServerCall.connection + + + + + + + + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + +Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/class-use/SimpleRpcServer.Connection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/SimpleRpcServer.Connection.html b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/SimpleRpcServer.Connection.html deleted file mode 100644 index 53f63b9..000 --- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/SimpleRpcServer.Connection.html +++ /dev/null @@ -1,248 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - - - - -Uses of Class org.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection (Apache HBase 2.0.0-SNAPSHOT API) - - - - - - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -Prev -Next - - -Frames -NoFrames - - -AllClasses - - - - - - - - - - -Uses of Classorg.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection - - - - - -Packages that use SimpleRpcServer.Connection - -Package -Description - - - -org.apache.hadoop.hbase.ipc - -Tools to help define network clients and servers. - - - - - - - - - - -Uses of SimpleRpcServer.Connection in org.apache.hadoop.hbase.ipc - -Fields in org.apache.hadoop.hbase.ipc with type parameters of type SimpleRpcServer.Connection - -Modifier and Type -Field and Description - - - -private
[08/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/nio/SingleByteBuff.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/nio/SingleByteBuff.html b/devapidocs/src-html/org/apache/hadoop/hbase/nio/SingleByteBuff.html index c44493b..0d816f9 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/nio/SingleByteBuff.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/nio/SingleByteBuff.html @@ -25,17 +25,17 @@ 017 */ 018package org.apache.hadoop.hbase.nio; 019 -020import java.io.IOException; -021import java.nio.ByteBuffer; -022import java.nio.channels.ReadableByteChannel; -023 -024import org.apache.hadoop.hbase.classification.InterfaceAudience; -025import org.apache.hadoop.hbase.util.ByteBufferUtils; -026import org.apache.hadoop.hbase.util.ObjectIntPair; -027import org.apache.hadoop.hbase.util.UnsafeAccess; -028import org.apache.hadoop.hbase.util.UnsafeAvailChecker; -029 -030import com.google.common.annotations.VisibleForTesting; +020import com.google.common.annotations.VisibleForTesting; +021 +022import java.io.IOException; +023import java.nio.ByteBuffer; +024import java.nio.channels.ReadableByteChannel; +025 +026import org.apache.hadoop.hbase.classification.InterfaceAudience; +027import org.apache.hadoop.hbase.util.ByteBufferUtils; +028import org.apache.hadoop.hbase.util.ObjectIntPair; +029import org.apache.hadoop.hbase.util.UnsafeAccess; +030import org.apache.hadoop.hbase.util.UnsafeAvailChecker; 031 032import sun.nio.ch.DirectBuffer; 033 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.SaslDigestCallbackHandler.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.SaslDigestCallbackHandler.html b/devapidocs/src-html/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.SaslDigestCallbackHandler.html index e144d00..2aea531 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.SaslDigestCallbackHandler.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.SaslDigestCallbackHandler.html @@ -31,25 +31,25 @@ 023import java.io.IOException; 024import java.util.Locale; 025import java.util.Map; -026 -027import javax.security.auth.callback.Callback; -028import javax.security.auth.callback.CallbackHandler; -029import javax.security.auth.callback.NameCallback; -030import javax.security.auth.callback.PasswordCallback; -031import javax.security.auth.callback.UnsupportedCallbackException; -032import javax.security.sasl.AuthorizeCallback; -033import javax.security.sasl.RealmCallback; -034 -035import org.apache.commons.logging.Log; -036import org.apache.commons.logging.LogFactory; -037import org.apache.hadoop.hbase.classification.InterfaceAudience; +026import java.util.function.Consumer; +027 +028import javax.security.auth.callback.Callback; +029import javax.security.auth.callback.CallbackHandler; +030import javax.security.auth.callback.NameCallback; +031import javax.security.auth.callback.PasswordCallback; +032import javax.security.auth.callback.UnsupportedCallbackException; +033import javax.security.sasl.AuthorizeCallback; +034import javax.security.sasl.RealmCallback; +035 +036import org.apache.commons.logging.Log; +037import org.apache.commons.logging.LogFactory; 038import org.apache.hadoop.conf.Configuration; -039import org.apache.hadoop.hbase.ipc.RpcServer; +039import org.apache.hadoop.hbase.classification.InterfaceAudience; 040import org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection; 041import org.apache.hadoop.security.UserGroupInformation; 042import org.apache.hadoop.security.token.SecretManager; -043import org.apache.hadoop.security.token.TokenIdentifier; -044import org.apache.hadoop.security.token.SecretManager.InvalidToken; +043import org.apache.hadoop.security.token.SecretManager.InvalidToken; +044import org.apache.hadoop.security.token.TokenIdentifier; 045 046/** 047 * A utility class for dealing with SASL on RPC server @@ -87,107 +87,105 @@ 079 /** CallbackHandler for SASL DIGEST-MD5 mechanism */ 080 public static class SaslDigestCallbackHandler implements CallbackHandler { 081private SecretManagerTokenIdentifier secretManager; -082private RpcServer.Connection connection; +082private ConsumerUserGroupInformation attemptingUserConsumer; 083 -084public SaslDigestCallbackHandler( -085 SecretManagerTokenIdentifier secretManager, -086RpcServer.Connection connection) { -087 this.secretManager = secretManager; -088 this.connection = connection; -089} -090 -091private char[] getPassword(TokenIdentifier tokenid) throws InvalidToken { -092 return
[36/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyServerRpcConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyServerRpcConnection.html b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyServerRpcConnection.html new file mode 100644 index 000..48fe4f7 --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/NettyServerRpcConnection.html @@ -0,0 +1,209 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +Uses of Class org.apache.hadoop.hbase.ipc.NettyServerRpcConnection (Apache HBase 2.0.0-SNAPSHOT API) + + + + + + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + + +Uses of Classorg.apache.hadoop.hbase.ipc.NettyServerRpcConnection + + + + + +Packages that use NettyServerRpcConnection + +Package +Description + + + +org.apache.hadoop.hbase.ipc + +Tools to help define network clients and servers. + + + + + + + + + + +Uses of NettyServerRpcConnection in org.apache.hadoop.hbase.ipc + +Fields in org.apache.hadoop.hbase.ipc declared as NettyServerRpcConnection + +Modifier and Type +Field and Description + + + +private NettyServerRpcConnection +NettyRpcServer.ConnectionHeaderHandler.connection + + +private NettyServerRpcConnection +NettyRpcServer.MessageDecoder.connection + + + + +Methods in org.apache.hadoop.hbase.ipc with parameters of type NettyServerRpcConnection + +Modifier and Type +Method and Description + + + +(package private) void +NettyRpcServer.MessageDecoder.setConnection(NettyServerRpcConnectionconnection) + + + + +Constructors in org.apache.hadoop.hbase.ipc with parameters of type NettyServerRpcConnection + +Constructor and Description + + + +NettyServerCall(intid, + org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingServiceservice, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptormd, + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeaderheader, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Messageparam, + CellScannercellScanner, + NettyServerRpcConnectionconnection, + longsize, + org.apache.htrace.TraceInfotinfo, + http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddressremoteAddress, + longreceiveTime, + inttimeout, + ByteBufferPoolreservoir, + CellBlockBuildercellBlockBuilder, + RpcServer.CallCleanupreqCleanup) + + + + + + + + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + +Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCall.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCall.html b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCall.html index a8aaa86..4c8c453 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCall.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCall.html @@ -113,7 +113,7 @@ (package private) class -ServerCall +ServerCallT extends ServerRpcConnection Datastructure that holds all necessary to a method invocation and then afterward, carries the result. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallContext.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallContext.html b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallContext.html index caf0355..4e0f0d4 100644 ---
[17/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerRpcConnection.ByteBuffByteInput.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerRpcConnection.ByteBuffByteInput.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerRpcConnection.ByteBuffByteInput.html new file mode 100644 index 000..50c4422 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerRpcConnection.ByteBuffByteInput.html @@ -0,0 +1,924 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.ipc; +019 +020import java.io.ByteArrayInputStream; +021import java.io.ByteArrayOutputStream; +022import java.io.Closeable; +023import java.io.DataOutputStream; +024import java.io.IOException; +025import java.net.InetAddress; +026import java.net.InetSocketAddress; +027import java.nio.ByteBuffer; +028import java.nio.channels.Channels; +029import java.nio.channels.ReadableByteChannel; +030import java.security.GeneralSecurityException; +031import java.security.PrivilegedExceptionAction; +032import java.util.Properties; +033 +034import javax.security.sasl.Sasl; +035import javax.security.sasl.SaslException; +036import javax.security.sasl.SaslServer; +037 +038import org.apache.commons.crypto.cipher.CryptoCipherFactory; +039import org.apache.commons.crypto.random.CryptoRandom; +040import org.apache.commons.crypto.random.CryptoRandomFactory; +041import org.apache.hadoop.hbase.CellScanner; +042import org.apache.hadoop.hbase.DoNotRetryIOException; +043import org.apache.hadoop.hbase.client.VersionInfoUtil; +044import org.apache.hadoop.hbase.codec.Codec; +045import org.apache.hadoop.hbase.io.ByteBufferOutputStream; +046import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; +047import org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup; +048import org.apache.hadoop.hbase.nio.ByteBuff; +049import org.apache.hadoop.hbase.nio.SingleByteBuff; +050import org.apache.hadoop.hbase.security.AccessDeniedException; +051import org.apache.hadoop.hbase.security.AuthMethod; +052import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; +053import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; +054import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; +055import org.apache.hadoop.hbase.security.SaslStatus; +056import org.apache.hadoop.hbase.security.SaslUtil; +057import org.apache.hadoop.hbase.security.User; +058import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; +059import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; +060import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; +061import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; +062import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +063import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; +064import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; +065import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; +066import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; +067import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; +068import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; +069import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; +070import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; +071import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation; +072import org.apache.hadoop.hbase.util.Bytes; +073import org.apache.hadoop.io.BytesWritable; +074import org.apache.hadoop.io.Writable; +075import org.apache.hadoop.io.WritableUtils; +076import org.apache.hadoop.io.compress.CompressionCodec; +077import org.apache.hadoop.security.UserGroupInformation; +078import org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod;
hbase-site git commit: INFRA-10751 Empty commit
Repository: hbase-site Updated Branches: refs/heads/asf-site f55ebeaa5 -> 8acdac3ac INFRA-10751 Empty commit Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/8acdac3a Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/8acdac3a Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/8acdac3a Branch: refs/heads/asf-site Commit: 8acdac3ac50e58b1e65ba09f745e36deb78694b3 Parents: f55ebea Author: jenkinsAuthored: Mon May 15 14:58:55 2017 + Committer: jenkins Committed: Mon May 15 14:58:55 2017 + -- --
[47/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html index 441b1a0..85c6f99 100644 --- a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html @@ -166,9 +166,9 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase -org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand org.apache.hadoop.hbase.backup.BackupType +org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand +org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase org.apache.hadoop.hbase.backup.BackupInfo.BackupState http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html b/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html index 3fa62ce..aa4c86f 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/CellScanner.html @@ -680,14 +680,13 @@ inttimeout) -abstract ServerCall -RpcServer.Connection.createCall(intid, +abstract ServerCall? +ServerRpcConnection.createCall(intid, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingServiceservice, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptormd, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeaderheader, org.apache.hadoop.hbase.shaded.com.google.protobuf.Messageparam, CellScannercellScanner, - RpcServer.Connectionconnection, longsize, org.apache.htrace.TraceInfotinfo, http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddressremoteAddress, @@ -695,14 +694,13 @@ RpcServer.CallCleanupreqCleanup) -ServerCall -SimpleRpcServer.Connection.createCall(intid, +NettyServerCall +NettyServerRpcConnection.createCall(intid, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingServiceservice, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptormd, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeaderheader, org.apache.hadoop.hbase.shaded.com.google.protobuf.Messageparam, CellScannercellScanner, - RpcServer.Connectionconnection, longsize, org.apache.htrace.TraceInfotinfo, http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddressremoteAddress, @@ -710,14 +708,13 @@ RpcServer.CallCleanupreqCleanup) -ServerCall -NettyRpcServer.NettyConnection.createCall(intid, +SimpleServerCall +SimpleServerRpcConnection.createCall(intid, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingServiceservice, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptormd, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeaderheader, org.apache.hadoop.hbase.shaded.com.google.protobuf.Messageparam, CellScannercellScanner, - RpcServer.Connectionconnection, longsize, org.apache.htrace.TraceInfotinfo, http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddressremoteAddress, @@ -811,13 +808,13 @@ HBaseRpcControllerImpl(CellScannercellScanner) -NettyServerCall(intid, +NettyServerCall(intid, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingServiceservice, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptormd, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeaderheader, org.apache.hadoop.hbase.shaded.com.google.protobuf.Messageparam, CellScannercellScanner, - RpcServer.Connectionconnection, + NettyServerRpcConnectionconnection, longsize,
[39/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html index 5c85b2d..f767071 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -private class SimpleRpcServer.Listener +private class SimpleRpcServer.Listener extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true; title="class or interface in java.lang">Thread Listens on the socket. Creates jobs for the handler threads @@ -296,7 +296,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? acceptChannel -privatehttp://docs.oracle.com/javase/8/docs/api/java/nio/channels/ServerSocketChannel.html?is-external=true; title="class or interface in java.nio.channels">ServerSocketChannel acceptChannel +privatehttp://docs.oracle.com/javase/8/docs/api/java/nio/channels/ServerSocketChannel.html?is-external=true; title="class or interface in java.nio.channels">ServerSocketChannel acceptChannel @@ -305,7 +305,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? selector -privatehttp://docs.oracle.com/javase/8/docs/api/java/nio/channels/Selector.html?is-external=true; title="class or interface in java.nio.channels">Selector selector +privatehttp://docs.oracle.com/javase/8/docs/api/java/nio/channels/Selector.html?is-external=true; title="class or interface in java.nio.channels">Selector selector @@ -314,7 +314,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? readers -privateSimpleRpcServer.Listener.Reader[] readers +privateSimpleRpcServer.Listener.Reader[] readers @@ -323,7 +323,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? currentReader -privateint currentReader +privateint currentReader @@ -332,7 +332,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? readerPendingConnectionQueueLength -private finalint readerPendingConnectionQueueLength +private finalint readerPendingConnectionQueueLength @@ -341,7 +341,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? readPool -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorService readPool +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorService readPool @@ -358,7 +358,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? Listener -publicListener(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) +publicListener(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -380,7 +380,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? run -publicvoidrun() +publicvoidrun() Specified by: http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--; title="class or interface in java.lang">runin interfacehttp://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true; title="class or interface in java.lang">Runnable @@ -395,7 +395,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? closeCurrentConnection -privatevoidcloseCurrentConnection(http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SelectionKey.html?is-external=true; title="class or interface in java.nio.channels">SelectionKeykey, +privatevoidcloseCurrentConnection(http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SelectionKey.html?is-external=true; title="class or interface in java.nio.channels">SelectionKeykey, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablee) @@ -405,7 +405,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html? getAddress -http://docs.oracle.com/javase/8/docs/api/java/net/InetSocketAddress.html?is-external=true; title="class or interface in java.net">InetSocketAddressgetAddress() +http://docs.oracle.com/javase/8/docs/api/java/net/InetSocketAddress.html?is-external=true;
[20/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html deleted file mode 100644 index 9e1c66c..000 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html +++ /dev/null @@ -1,1689 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - -Source code - - - - -001/** -002 * Licensed to the Apache Software Foundation (ASF) under one -003 * or more contributor license agreements. See the NOTICE file -004 * distributed with this work for additional information -005 * regarding copyright ownership. The ASF licenses this file -006 * to you under the Apache License, Version 2.0 (the -007 * "License"); you may not use this file except in compliance -008 * with the License. You may obtain a copy of the License at -009 * -010 * http://www.apache.org/licenses/LICENSE-2.0 -011 * -012 * Unless required by applicable law or agreed to in writing, software -013 * distributed under the License is distributed on an "AS IS" BASIS, -014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -015 * See the License for the specific language governing permissions and -016 * limitations under the License. -017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHORIZATION; -022 -023import java.io.ByteArrayInputStream; -024import java.io.ByteArrayOutputStream; -025import java.io.Closeable; -026import java.io.DataOutputStream; -027import java.io.IOException; -028import java.net.InetAddress; -029import java.net.InetSocketAddress; -030import java.nio.ByteBuffer; -031import java.nio.channels.Channels; -032import java.nio.channels.GatheringByteChannel; -033import java.nio.channels.ReadableByteChannel; -034import java.nio.channels.WritableByteChannel; -035import java.security.GeneralSecurityException; -036import java.security.PrivilegedExceptionAction; -037import java.util.ArrayList; -038import java.util.HashMap; -039import java.util.List; -040import java.util.Map; -041import java.util.Properties; -042import java.util.concurrent.atomic.LongAdder; -043 -044import javax.security.sasl.Sasl; -045import javax.security.sasl.SaslException; -046import javax.security.sasl.SaslServer; -047 -048import org.apache.commons.crypto.cipher.CryptoCipherFactory; -049import org.apache.commons.crypto.random.CryptoRandom; -050import org.apache.commons.crypto.random.CryptoRandomFactory; -051import org.apache.commons.logging.Log; -052import org.apache.commons.logging.LogFactory; -053import org.apache.hadoop.conf.Configuration; -054import org.apache.hadoop.hbase.CallQueueTooBigException; -055import org.apache.hadoop.hbase.CellScanner; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseInterfaceAudience; -058import org.apache.hadoop.hbase.HConstants; -059import org.apache.hadoop.hbase.Server; -060import org.apache.hadoop.hbase.classification.InterfaceAudience; -061import org.apache.hadoop.hbase.classification.InterfaceStability; -062import org.apache.hadoop.hbase.client.VersionInfoUtil; -063import org.apache.hadoop.hbase.codec.Codec; -064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -066import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -067import org.apache.hadoop.hbase.io.ByteBufferPool; -068import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -069import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -070import org.apache.hadoop.hbase.monitoring.TaskMonitor; -071import org.apache.hadoop.hbase.nio.ByteBuff; -072import org.apache.hadoop.hbase.nio.MultiByteBuff; -073import org.apache.hadoop.hbase.nio.SingleByteBuff; -074import org.apache.hadoop.hbase.regionserver.RSRpcServices; -075import org.apache.hadoop.hbase.security.AccessDeniedException; -076import org.apache.hadoop.hbase.security.AuthMethod; -077import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -078import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -080import org.apache.hadoop.hbase.security.SaslStatus; -081import org.apache.hadoop.hbase.security.SaslUtil; -082import org.apache.hadoop.hbase.security.User; -083import org.apache.hadoop.hbase.security.UserProvider; -084import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -085import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -086import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; -087import
[04/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-archetypes/hbase-client-project/dependency-info.html -- diff --git a/hbase-archetypes/hbase-client-project/dependency-info.html b/hbase-archetypes/hbase-client-project/dependency-info.html index 6bb94d1..73faf31 100644 --- a/hbase-archetypes/hbase-client-project/dependency-info.html +++ b/hbase-archetypes/hbase-client-project/dependency-info.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15 | Version: 2.0.0-SNAPSHOT Apache HBase - Exemplar for hbase-client archetype http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-archetypes/hbase-client-project/dependency-management.html -- diff --git a/hbase-archetypes/hbase-client-project/dependency-management.html b/hbase-archetypes/hbase-client-project/dependency-management.html index 7846215..581f8f9 100644 --- a/hbase-archetypes/hbase-client-project/dependency-management.html +++ b/hbase-archetypes/hbase-client-project/dependency-management.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15 | Version: 2.0.0-SNAPSHOT Apache HBase - Exemplar for hbase-client archetype http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-archetypes/hbase-client-project/index.html -- diff --git a/hbase-archetypes/hbase-client-project/index.html b/hbase-archetypes/hbase-client-project/index.html index fb2784b..09145c0 100644 --- a/hbase-archetypes/hbase-client-project/index.html +++ b/hbase-archetypes/hbase-client-project/index.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15 | Version: 2.0.0-SNAPSHOT Apache HBase - Exemplar for hbase-client archetype http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-archetypes/hbase-client-project/integration.html -- diff --git a/hbase-archetypes/hbase-client-project/integration.html b/hbase-archetypes/hbase-client-project/integration.html index 1584d08..90d0298 100644 --- a/hbase-archetypes/hbase-client-project/integration.html +++ b/hbase-archetypes/hbase-client-project/integration.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15 | Version: 2.0.0-SNAPSHOT Apache HBase - Exemplar for hbase-client archetype http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-archetypes/hbase-client-project/issue-tracking.html -- diff --git a/hbase-archetypes/hbase-client-project/issue-tracking.html b/hbase-archetypes/hbase-client-project/issue-tracking.html index e9e106f..27217f1 100644 --- a/hbase-archetypes/hbase-client-project/issue-tracking.html +++ b/hbase-archetypes/hbase-client-project/issue-tracking.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15
[50/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/checkstyle-aggregate.html -- diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html index 52db34f..470aa62 100644 --- a/checkstyle-aggregate.html +++ b/checkstyle-aggregate.html @@ -7,7 +7,7 @@ - + Apache HBase Checkstyle Results @@ -286,10 +286,10 @@ Warnings Errors -2161 +2165 0 0 -14359 +14353 Files @@ -2459,10 +2459,10 @@ 0 1 -org/apache/hadoop/hbase/ipc/NettyRpcServer.java +org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.java 0 0 -3 +1 org/apache/hadoop/hbase/ipc/PriorityFunction.java 0 @@ -2522,7 +2522,7 @@ org/apache/hadoop/hbase/ipc/RpcServer.java 0 0 -53 +11 org/apache/hadoop/hbase/ipc/RpcServerFactory.java 0 @@ -2539,15 +2539,35 @@ 0 3 +org/apache/hadoop/hbase/ipc/ServerRpcConnection.java +0 +0 +45 + org/apache/hadoop/hbase/ipc/SimpleRpcScheduler.java 0 0 8 - + org/apache/hadoop/hbase/ipc/SimpleRpcServer.java 0 0 -26 +13 + +org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.java +0 +0 +4 + +org/apache/hadoop/hbase/ipc/SimpleServerCall.java +0 +0 +1 + +org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.java +0 +0 +4 org/apache/hadoop/hbase/mapred/Driver.java 0 @@ -3547,7 +3567,7 @@ org/apache/hadoop/hbase/nio/SingleByteBuff.java 0 0 -2 +1 org/apache/hadoop/hbase/procedure/MasterProcedureManager.java 0 @@ -5332,7 +5352,7 @@ org/apache/hadoop/hbase/security/HBaseSaslRpcServer.java 0 0 -4 +2 org/apache/hadoop/hbase/security/SaslStatus.java 0 @@ -7079,7 +7099,7 @@ http://checkstyle.sourceforge.net/config_blocks.html#LeftCurly;>LeftCurly -355 +353 Error @@ -7132,7 +7152,7 @@ protectedAllowed: true packageAllowed: true -140 +139 Error imports @@ -7146,7 +7166,7 @@ ordered: true sortStaticImportsAlphabetically: true option: top -917 +915 Error @@ -7169,19 +7189,19 @@ caseIndent: 2 basicOffset: 2 lineWrappingIndentation: 2 -5076 +5072 Error javadoc http://checkstyle.sourceforge.net/config_javadoc.html#JavadocTagContinuationIndentation;>JavadocTagContinuationIndentation offset: 2 -787 +780 Error http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription;>NonEmptyAtclauseDescription -3264 +3269 Error misc @@ -7199,7 +7219,7 @@ max: 100 ignorePattern: ^package.*|^import.*|a href|href|http://|https://|ftp://|org.apache.thrift.|com.google.protobuf.|hbase.protobuf.generated -773 +778 Error @@ -12875,7 +12895,7 @@ Error javadoc -JavadocTagContinuationIndentation +NonEmptyAtclauseDescription Javadoc comment at column 43 has parse error. Missed HTML close tag 'TableName'. Sometimes it means that close tag missed for one of previous tags. 123 @@ -15953,7 +15973,7 @@ Error javadoc -JavadocTagContinuationIndentation +NonEmptyAtclauseDescription Javadoc comment at column 64 has parse error. Missed HTML close tag 'code'. Sometimes it means that close tag missed for one of previous tags. 1957 @@ -20501,7 +20521,7 @@ Error javadoc -JavadocTagContinuationIndentation +NonEmptyAtclauseDescription Javadoc comment at column 37 has parse error. Details: no viable alternative at input 'ColumnFamily,' while parsing HTML_ELEMENT 29 @@ -33546,7 +33566,7 @@ Wrong order for 'io.netty.buffer.ByteBuf' import. 24 -org/apache/hadoop/hbase/ipc/NettyRpcServer.java +org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.java Severity @@ -33556,22 +33576,10 @@ Line Error -indentation -Indentation -'array initialization lcurly' have incorrect indentation level 10, expected level should be one of the following: 6, 8. -217 - -Error -indentation -Indentation -'if' child have incorrect indentation level 9, expected level should be 8. -219 - -Error -indentation -Indentation -'if' child have incorrect indentation level 9, expected level should be 8. -222 +imports +ImportOrder +Wrong order for 'org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor' import. +42 org/apache/hadoop/hbase/ipc/PriorityFunction.java @@ -33880,322 +33888,70 @@ Line Error -imports -ImportOrder -Wrong order for 'com.google.common.annotations.VisibleForTesting' import. -119 - -Error -design -VisibilityModifier -Variable 'attemptingUser' must be private and have accessor methods. -322 - -Error -indentation -Indentation -'method def' child have incorrect indentation level 6, expected level should be 8. -356 - -Error -indentation -Indentation -'method def' child have incorrect indentation level 6, expected level should be 8. -357 - -Error -javadoc -NonEmptyAtclauseDescription -At-clause should have a non-empty description. -381 - -Error -blocks -NeedBraces -'if' construct must use '{}'s. -386 - -Error -blocks -NeedBraces -'if' construct must use '{}'s. -388 - -Error -blocks -NeedBraces -'if' construct must use '{}'s. -394 - -Error -javadoc -NonEmptyAtclauseDescription
[49/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index 3ffa35e..775c114 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -25,8 +25,8 @@ under the License. en-us 2007 - 2017 The Apache Software Foundation - File: 2161, - Errors: 14359, + File: 2165, + Errors: 14353, Warnings: 0, Infos: 0 @@ -7620,6 +7620,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.SimpleRpcServerResponder.java;>org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.java + + + 0 + + + 0 + + + 4 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.zookeeper.RecoveringRegionWatcher.java;>org/apache/hadoop/hbase/zookeeper/RecoveringRegionWatcher.java @@ -9482,6 +9496,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.ServerRpcConnection.java;>org/apache/hadoop/hbase/ipc/ServerRpcConnection.java + + + 0 + + + 0 + + + 45 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.client.ScanResultConsumer.java;>org/apache/hadoop/hbase/client/ScanResultConsumer.java @@ -9832,6 +9860,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.SimpleServerRpcConnection.java;>org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.java + + + 0 + + + 0 + + + 4 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.compactions.CompactionWindowFactory.java;>org/apache/hadoop/hbase/regionserver/compactions/CompactionWindowFactory.java @@ -11801,7 +11843,7 @@ under the License. 0 - 26 + 13 @@ -15847,7 +15889,7 @@ under the License. 0 - 0 + 1 @@ -17009,7 +17051,7 @@ under the License. 0 - 53 + 11 @@ -20313,7 +20355,7 @@ under the License. 0 - 3 + 0 @@ -20803,7 +20845,7 @@ under the License. 0 - 2 + 1 @@ -26389,7 +26431,7 @@ under the License. 0 - 4 + 2 @@ -26520,6 +26562,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.ipc.NettyServerRpcConnection.java;>org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.java + + + 0 + + + 0 + + + 1 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.io.util.LRUDictionary.java;>org/apache/hadoop/hbase/io/util/LRUDictionary.java http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/coc.html -- diff --git a/coc.html b/coc.html index a11a43f..efdf73e 100644 --- a/coc.html +++ b/coc.html @@ -7,7 +7,7
[11/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html deleted file mode 100644 index 7f61b54..000 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Responder.html +++ /dev/null @@ -1,1446 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - -Source code - - - - -001/** -002 * Licensed to the Apache Software Foundation (ASF) under one -003 * or more contributor license agreements. See the NOTICE file -004 * distributed with this work for additional information -005 * regarding copyright ownership. The ASF licenses this file -006 * to you under the Apache License, Version 2.0 (the -007 * "License"); you may not use this file except in compliance -008 * with the License. You may obtain a copy of the License at -009 * -010 * http://www.apache.org/licenses/LICENSE-2.0 -011 * -012 * Unless required by applicable law or agreed to in writing, software -013 * distributed under the License is distributed on an "AS IS" BASIS, -014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -015 * See the License for the specific language governing permissions and -016 * limitations under the License. -017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import java.io.IOException; -022import java.io.InputStream; -023import java.net.BindException; -024import java.net.InetAddress; -025import java.net.InetSocketAddress; -026import java.net.ServerSocket; -027import java.net.Socket; -028import java.net.SocketException; -029import java.net.UnknownHostException; -030import java.nio.ByteBuffer; -031import java.nio.channels.CancelledKeyException; -032import java.nio.channels.ClosedChannelException; -033import java.nio.channels.GatheringByteChannel; -034import java.nio.channels.ReadableByteChannel; -035import java.nio.channels.SelectionKey; -036import java.nio.channels.Selector; -037import java.nio.channels.ServerSocketChannel; -038import java.nio.channels.SocketChannel; -039import java.util.ArrayList; -040import java.util.Arrays; -041import java.util.Collections; -042import java.util.Iterator; -043import java.util.List; -044import java.util.Set; -045import java.util.Timer; -046import java.util.TimerTask; -047import java.util.concurrent.ConcurrentHashMap; -048import java.util.concurrent.ConcurrentLinkedDeque; -049import java.util.concurrent.ExecutorService; -050import java.util.concurrent.Executors; -051import java.util.concurrent.LinkedBlockingQueue; -052import java.util.concurrent.atomic.AtomicInteger; -053import java.util.concurrent.atomic.LongAdder; -054import java.util.concurrent.locks.Lock; -055import java.util.concurrent.locks.ReentrantLock; -056 -057import org.apache.hadoop.conf.Configuration; -058import org.apache.hadoop.hbase.CellScanner; -059import org.apache.hadoop.hbase.DoNotRetryIOException; -060import org.apache.hadoop.hbase.HBaseIOException; -061import org.apache.hadoop.hbase.HConstants; -062import org.apache.hadoop.hbase.Server; -063import org.apache.hadoop.hbase.classification.InterfaceAudience; -064import org.apache.hadoop.hbase.classification.InterfaceStability; -065import org.apache.hadoop.hbase.client.VersionInfoUtil; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -068import org.apache.hadoop.hbase.nio.ByteBuff; -069import org.apache.hadoop.hbase.nio.SingleByteBuff; -070import org.apache.hadoop.hbase.security.AccessDeniedException; -071import org.apache.hadoop.hbase.security.AuthMethod; -072import org.apache.hadoop.hbase.security.HBasePolicyProvider; -073import org.apache.hadoop.hbase.security.SaslStatus; -074import org.apache.hadoop.hbase.security.SaslUtil; -075import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -076import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -077import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -078import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -079import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -081import org.apache.hadoop.hbase.util.Bytes; -082import org.apache.hadoop.hbase.util.Pair; -083import org.apache.hadoop.hbase.util.Threads; -084import org.apache.hadoop.io.IOUtils; -085import org.apache.hadoop.io.IntWritable; -086import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -087import org.apache.hadoop.util.StringUtils; -088import org.apache.htrace.TraceInfo; -089 -090import com.google.common.util.concurrent.ThreadFactoryBuilder;
[19/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html index 9e1c66c..27ccbd5 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.html @@ -28,1601 +28,787 @@ 020 021import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHORIZATION; 022 -023import java.io.ByteArrayInputStream; -024import java.io.ByteArrayOutputStream; -025import java.io.Closeable; -026import java.io.DataOutputStream; -027import java.io.IOException; -028import java.net.InetAddress; -029import java.net.InetSocketAddress; -030import java.nio.ByteBuffer; -031import java.nio.channels.Channels; -032import java.nio.channels.GatheringByteChannel; -033import java.nio.channels.ReadableByteChannel; -034import java.nio.channels.WritableByteChannel; -035import java.security.GeneralSecurityException; -036import java.security.PrivilegedExceptionAction; -037import java.util.ArrayList; -038import java.util.HashMap; -039import java.util.List; -040import java.util.Map; -041import java.util.Properties; -042import java.util.concurrent.atomic.LongAdder; -043 -044import javax.security.sasl.Sasl; -045import javax.security.sasl.SaslException; -046import javax.security.sasl.SaslServer; -047 -048import org.apache.commons.crypto.cipher.CryptoCipherFactory; -049import org.apache.commons.crypto.random.CryptoRandom; -050import org.apache.commons.crypto.random.CryptoRandomFactory; -051import org.apache.commons.logging.Log; -052import org.apache.commons.logging.LogFactory; -053import org.apache.hadoop.conf.Configuration; -054import org.apache.hadoop.hbase.CallQueueTooBigException; -055import org.apache.hadoop.hbase.CellScanner; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseInterfaceAudience; -058import org.apache.hadoop.hbase.HConstants; -059import org.apache.hadoop.hbase.Server; -060import org.apache.hadoop.hbase.classification.InterfaceAudience; -061import org.apache.hadoop.hbase.classification.InterfaceStability; -062import org.apache.hadoop.hbase.client.VersionInfoUtil; -063import org.apache.hadoop.hbase.codec.Codec; -064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -066import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -067import org.apache.hadoop.hbase.io.ByteBufferPool; -068import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -069import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -070import org.apache.hadoop.hbase.monitoring.TaskMonitor; -071import org.apache.hadoop.hbase.nio.ByteBuff; -072import org.apache.hadoop.hbase.nio.MultiByteBuff; -073import org.apache.hadoop.hbase.nio.SingleByteBuff; -074import org.apache.hadoop.hbase.regionserver.RSRpcServices; -075import org.apache.hadoop.hbase.security.AccessDeniedException; -076import org.apache.hadoop.hbase.security.AuthMethod; -077import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -078import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -080import org.apache.hadoop.hbase.security.SaslStatus; -081import org.apache.hadoop.hbase.security.SaslUtil; -082import org.apache.hadoop.hbase.security.User; -083import org.apache.hadoop.hbase.security.UserProvider; -084import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -085import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -086import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; -087import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; -088import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -089import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -090import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -091import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; -092import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; -093import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; -094import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; -099import
[23/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html index 9e1c66c..27ccbd5 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html @@ -28,1601 +28,787 @@ 020 021import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHORIZATION; 022 -023import java.io.ByteArrayInputStream; -024import java.io.ByteArrayOutputStream; -025import java.io.Closeable; -026import java.io.DataOutputStream; -027import java.io.IOException; -028import java.net.InetAddress; -029import java.net.InetSocketAddress; -030import java.nio.ByteBuffer; -031import java.nio.channels.Channels; -032import java.nio.channels.GatheringByteChannel; -033import java.nio.channels.ReadableByteChannel; -034import java.nio.channels.WritableByteChannel; -035import java.security.GeneralSecurityException; -036import java.security.PrivilegedExceptionAction; -037import java.util.ArrayList; -038import java.util.HashMap; -039import java.util.List; -040import java.util.Map; -041import java.util.Properties; -042import java.util.concurrent.atomic.LongAdder; -043 -044import javax.security.sasl.Sasl; -045import javax.security.sasl.SaslException; -046import javax.security.sasl.SaslServer; -047 -048import org.apache.commons.crypto.cipher.CryptoCipherFactory; -049import org.apache.commons.crypto.random.CryptoRandom; -050import org.apache.commons.crypto.random.CryptoRandomFactory; -051import org.apache.commons.logging.Log; -052import org.apache.commons.logging.LogFactory; -053import org.apache.hadoop.conf.Configuration; -054import org.apache.hadoop.hbase.CallQueueTooBigException; -055import org.apache.hadoop.hbase.CellScanner; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseInterfaceAudience; -058import org.apache.hadoop.hbase.HConstants; -059import org.apache.hadoop.hbase.Server; -060import org.apache.hadoop.hbase.classification.InterfaceAudience; -061import org.apache.hadoop.hbase.classification.InterfaceStability; -062import org.apache.hadoop.hbase.client.VersionInfoUtil; -063import org.apache.hadoop.hbase.codec.Codec; -064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -066import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -067import org.apache.hadoop.hbase.io.ByteBufferPool; -068import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -069import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -070import org.apache.hadoop.hbase.monitoring.TaskMonitor; -071import org.apache.hadoop.hbase.nio.ByteBuff; -072import org.apache.hadoop.hbase.nio.MultiByteBuff; -073import org.apache.hadoop.hbase.nio.SingleByteBuff; -074import org.apache.hadoop.hbase.regionserver.RSRpcServices; -075import org.apache.hadoop.hbase.security.AccessDeniedException; -076import org.apache.hadoop.hbase.security.AuthMethod; -077import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -078import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -080import org.apache.hadoop.hbase.security.SaslStatus; -081import org.apache.hadoop.hbase.security.SaslUtil; -082import org.apache.hadoop.hbase.security.User; -083import org.apache.hadoop.hbase.security.UserProvider; -084import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -085import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -086import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; -087import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; -088import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -089import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -090import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -091import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; -092import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; -093import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; -094import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; -098import
[27/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html index 109b5f3..e484176 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageEncoder.html @@ -23,484 +23,310 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import io.netty.bootstrap.ServerBootstrap; -022import io.netty.buffer.ByteBuf; -023import io.netty.buffer.PooledByteBufAllocator; -024import io.netty.buffer.Unpooled; -025import io.netty.channel.Channel; -026import io.netty.channel.ChannelFuture; -027import io.netty.channel.ChannelFutureListener; -028import io.netty.channel.ChannelHandlerContext; -029import io.netty.channel.ChannelInboundHandlerAdapter; -030import io.netty.channel.ChannelInitializer; -031import io.netty.channel.ChannelOption; -032import io.netty.channel.ChannelOutboundHandlerAdapter; -033import io.netty.channel.ChannelPipeline; -034import io.netty.channel.ChannelPromise; -035import io.netty.channel.EventLoopGroup; -036import io.netty.channel.epoll.EpollEventLoopGroup; -037import io.netty.channel.epoll.EpollServerSocketChannel; -038import io.netty.channel.group.ChannelGroup; -039import io.netty.channel.group.DefaultChannelGroup; -040import io.netty.channel.nio.NioEventLoopGroup; -041import io.netty.channel.socket.SocketChannel; -042import io.netty.channel.socket.nio.NioServerSocketChannel; -043import io.netty.handler.codec.ByteToMessageDecoder; -044import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -045import io.netty.util.concurrent.GlobalEventExecutor; -046 -047import java.io.IOException; -048import java.io.InterruptedIOException; -049import java.net.InetAddress; -050import java.net.InetSocketAddress; -051import java.nio.ByteBuffer; -052import java.util.Arrays; -053import java.util.List; -054import java.util.concurrent.CountDownLatch; -055 -056import org.apache.commons.logging.Log; -057import org.apache.commons.logging.LogFactory; -058import org.apache.hadoop.conf.Configuration; -059import org.apache.hadoop.hbase.CellScanner; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -063import org.apache.hadoop.hbase.nio.ByteBuff; -064import org.apache.hadoop.hbase.nio.SingleByteBuff; -065import org.apache.hadoop.hbase.security.AccessDeniedException; -066import org.apache.hadoop.hbase.security.AuthMethod; -067import org.apache.hadoop.hbase.security.HBasePolicyProvider; -068import org.apache.hadoop.hbase.security.SaslStatus; -069import org.apache.hadoop.hbase.security.SaslUtil; -070import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -073import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -074import org.apache.hadoop.hbase.util.Bytes; -075import org.apache.hadoop.hbase.util.JVM; -076import org.apache.hadoop.hbase.util.Pair; -077import org.apache.hadoop.io.IntWritable; -078import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -079import org.apache.htrace.TraceInfo; +018package org.apache.hadoop.hbase.ipc; +019 +020import io.netty.bootstrap.ServerBootstrap; +021import io.netty.buffer.ByteBuf; +022import io.netty.buffer.PooledByteBufAllocator; +023import io.netty.buffer.Unpooled; +024import io.netty.channel.Channel; +025import io.netty.channel.ChannelFuture; +026import io.netty.channel.ChannelFutureListener; +027import io.netty.channel.ChannelHandlerContext; +028import io.netty.channel.ChannelInboundHandlerAdapter; +029import io.netty.channel.ChannelInitializer; +030import io.netty.channel.ChannelOption; +031import io.netty.channel.ChannelOutboundHandlerAdapter; +032import io.netty.channel.ChannelPipeline; +033import io.netty.channel.ChannelPromise; +034import io.netty.channel.EventLoopGroup; +035import io.netty.channel.epoll.EpollEventLoopGroup; +036import io.netty.channel.epoll.EpollServerSocketChannel; +037import io.netty.channel.group.ChannelGroup; +038import io.netty.channel.group.DefaultChannelGroup; +039import io.netty.channel.nio.NioEventLoopGroup; +040import io.netty.channel.socket.SocketChannel; +041import io.netty.channel.socket.nio.NioServerSocketChannel; +042import io.netty.handler.codec.ByteToMessageDecoder; +043import
[48/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/index-all.html -- diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html index 471aa21..1779018 100644 --- a/devapidocs/index-all.html +++ b/devapidocs/index-all.html @@ -970,7 +970,7 @@ add(CallRunner) - Method in class org.apache.hadoop.hbase.ipc.AdaptiveLifoCoDelCallQueue -add(SimpleRpcServer.Connection) - Method in class org.apache.hadoop.hbase.ipc.SimpleRpcServer.ConnectionManager +add(SimpleServerRpcConnection) - Method in class org.apache.hadoop.hbase.ipc.SimpleRpcServer.ConnectionManager add(ServerName) - Method in class org.apache.hadoop.hbase.master.DeadServer @@ -1427,7 +1427,7 @@ addConfiguration(String, String) - Method in class org.apache.hadoop.hbase.NamespaceDescriptor.Builder -addConnection(SimpleRpcServer.Connection) - Method in class org.apache.hadoop.hbase.ipc.SimpleRpcServer.Listener.Reader +addConnection(SimpleServerRpcConnection) - Method in class org.apache.hadoop.hbase.ipc.SimpleRpcServer.Listener.Reader Updating the readSelector while it's being used is not thread-safe, so the connection must be queued. @@ -2065,7 +2065,7 @@ addr - Variable in class org.apache.hadoop.hbase.ipc.AbstractRpcClient.AbstractRpcChannel -addr - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Connection +addr - Variable in class org.apache.hadoop.hbase.ipc.ServerRpcConnection addReferenceFiles(SnapshotManifest.RegionVisitor, Object, Object, CollectionStoreFileInfo, boolean) - Method in class org.apache.hadoop.hbase.snapshot.SnapshotManifest @@ -4152,7 +4152,9 @@ Check if we should attain safe point. -attemptingUser - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Connection +attemptingUser - Variable in class org.apache.hadoop.hbase.ipc.ServerRpcConnection + +attemptingUserConsumer - Variable in class org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler attempts - Variable in class org.apache.hadoop.hbase.util.RetryCounter @@ -4190,7 +4192,7 @@ AUTH_TOKEN_TYPE - Static variable in class org.apache.hadoop.hbase.security.token.AuthenticationTokenIdentifier -authenticatedWithFallback - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Connection +authenticatedWithFallback - Variable in class org.apache.hadoop.hbase.ipc.ServerRpcConnection AUTHENTICATION_FAILURES_DESC - Static variable in interface org.apache.hadoop.hbase.ipc.MetricsHBaseServerSource @@ -4265,11 +4267,11 @@ AuthenticationTokenSelector() - Constructor for class org.apache.hadoop.hbase.security.token.AuthenticationTokenSelector -authFailedCall - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Connection +authFailedCall - Variable in class org.apache.hadoop.hbase.ipc.ServerRpcConnection authFailedFailedOpCount - Variable in class org.apache.hadoop.hbase.zookeeper.MetricsZooKeeperSourceImpl -authFailedResponse - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Connection +authFailedResponse - Variable in class org.apache.hadoop.hbase.ipc.ServerRpcConnection AuthFilter - Class in org.apache.hadoop.hbase.rest.filter @@ -4285,7 +4287,7 @@ authMethod - Variable in class org.apache.hadoop.hbase.ipc.RpcConnection -authMethod - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Connection +authMethod - Variable in class org.apache.hadoop.hbase.ipc.ServerRpcConnection AuthMethod - Enum in org.apache.hadoop.hbase.security @@ -4295,7 +4297,7 @@ AUTHORIZATION - Static variable in class org.apache.hadoop.hbase.thrift.ThriftHttpServlet -AUTHORIZATION_FAILED_CALLID - Static variable in class org.apache.hadoop.hbase.ipc.RpcServer.Connection +AUTHORIZATION_FAILED_CALLID - Static variable in class org.apache.hadoop.hbase.ipc.ServerRpcConnection AUTHORIZATION_FAILURES_DESC - Static variable in interface org.apache.hadoop.hbase.ipc.MetricsHBaseServerSource @@ -4369,7 +4371,7 @@ authorize(User, TableName, byte[], Permission.Action) - Method in class org.apache.hadoop.hbase.security.access.TableAuthManager -authorizeConnection() - Method in class org.apache.hadoop.hbase.ipc.RpcServer.Connection +authorizeConnection() - Method in class org.apache.hadoop.hbase.ipc.ServerRpcConnection authorizeGroup(String, Permission.Action) - Method in class org.apache.hadoop.hbase.security.access.TableAuthManager @@ -6660,7 +6662,7 @@ buf - Variable in class org.apache.hadoop.hbase.ipc.CellBlockBuilder.ByteBufOutputStreamSupplier -buf - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.ByteBuffByteInput +buf - Variable in class org.apache.hadoop.hbase.ipc.ServerRpcConnection.ByteBuffByteInput buf - Variable in class org.apache.hadoop.hbase.nio.SingleByteBuff @@ -7336,7 +7338,7 @@ ByteBuff() - Constructor for class org.apache.hadoop.hbase.nio.ByteBuff -ByteBuffByteInput(ByteBuff, int, int) - Constructor
[03/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-spark/mail-lists.html -- diff --git a/hbase-spark/mail-lists.html b/hbase-spark/mail-lists.html index 3b34be6..f32604c 100644 --- a/hbase-spark/mail-lists.html +++ b/hbase-spark/mail-lists.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15 | Version: 2.0.0-SNAPSHOT Apache HBase - Spark http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-spark/plugin-management.html -- diff --git a/hbase-spark/plugin-management.html b/hbase-spark/plugin-management.html index 7705526..d495230 100644 --- a/hbase-spark/plugin-management.html +++ b/hbase-spark/plugin-management.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15 | Version: 2.0.0-SNAPSHOT Apache HBase - Spark http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-spark/plugins.html -- diff --git a/hbase-spark/plugins.html b/hbase-spark/plugins.html index 41c48ef..91c67ca 100644 --- a/hbase-spark/plugins.html +++ b/hbase-spark/plugins.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15 | Version: 2.0.0-SNAPSHOT Apache HBase - Spark http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-spark/project-info.html -- diff --git a/hbase-spark/project-info.html b/hbase-spark/project-info.html index 61bebc4..a74281e 100644 --- a/hbase-spark/project-info.html +++ b/hbase-spark/project-info.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15 | Version: 2.0.0-SNAPSHOT Apache HBase - Spark http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-spark/project-reports.html -- diff --git a/hbase-spark/project-reports.html b/hbase-spark/project-reports.html index 14a56f0..5a451c0 100644 --- a/hbase-spark/project-reports.html +++ b/hbase-spark/project-reports.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -Last Published: 2017-05-14 +Last Published: 2017-05-15 | Version: 2.0.0-SNAPSHOT Apache HBase - Spark http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/hbase-spark/project-summary.html -- diff --git a/hbase-spark/project-summary.html b/hbase-spark/project-summary.html index e68bc91..73ce5d8 100644 --- a/hbase-spark/project-summary.html +++ b/hbase-spark/project-summary.html @@ -1,5 +1,5 @@ http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd;> - + http://www.w3.org/1999/xhtml; xml:lang="en" lang="en"> @@ -10,7 +10,7 @@ @import url("./css/site.css"); - + @@ -27,7 +27,7 @@ -
[13/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html index 7f61b54..35e8890 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Listener.Reader.html @@ -23,1363 +23,690 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import java.io.IOException; -022import java.io.InputStream; -023import java.net.BindException; -024import java.net.InetAddress; -025import java.net.InetSocketAddress; -026import java.net.ServerSocket; -027import java.net.Socket; -028import java.net.SocketException; -029import java.net.UnknownHostException; -030import java.nio.ByteBuffer; -031import java.nio.channels.CancelledKeyException; -032import java.nio.channels.ClosedChannelException; -033import java.nio.channels.GatheringByteChannel; -034import java.nio.channels.ReadableByteChannel; -035import java.nio.channels.SelectionKey; -036import java.nio.channels.Selector; -037import java.nio.channels.ServerSocketChannel; -038import java.nio.channels.SocketChannel; -039import java.util.ArrayList; -040import java.util.Arrays; -041import java.util.Collections; -042import java.util.Iterator; -043import java.util.List; -044import java.util.Set; -045import java.util.Timer; -046import java.util.TimerTask; -047import java.util.concurrent.ConcurrentHashMap; -048import java.util.concurrent.ConcurrentLinkedDeque; -049import java.util.concurrent.ExecutorService; -050import java.util.concurrent.Executors; -051import java.util.concurrent.LinkedBlockingQueue; -052import java.util.concurrent.atomic.AtomicInteger; -053import java.util.concurrent.atomic.LongAdder; -054import java.util.concurrent.locks.Lock; -055import java.util.concurrent.locks.ReentrantLock; -056 -057import org.apache.hadoop.conf.Configuration; -058import org.apache.hadoop.hbase.CellScanner; -059import org.apache.hadoop.hbase.DoNotRetryIOException; -060import org.apache.hadoop.hbase.HBaseIOException; -061import org.apache.hadoop.hbase.HConstants; -062import org.apache.hadoop.hbase.Server; -063import org.apache.hadoop.hbase.classification.InterfaceAudience; -064import org.apache.hadoop.hbase.classification.InterfaceStability; -065import org.apache.hadoop.hbase.client.VersionInfoUtil; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -068import org.apache.hadoop.hbase.nio.ByteBuff; -069import org.apache.hadoop.hbase.nio.SingleByteBuff; -070import org.apache.hadoop.hbase.security.AccessDeniedException; -071import org.apache.hadoop.hbase.security.AuthMethod; -072import org.apache.hadoop.hbase.security.HBasePolicyProvider; -073import org.apache.hadoop.hbase.security.SaslStatus; -074import org.apache.hadoop.hbase.security.SaslUtil; -075import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -076import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -077import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -078import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -079import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -081import org.apache.hadoop.hbase.util.Bytes; -082import org.apache.hadoop.hbase.util.Pair; -083import org.apache.hadoop.hbase.util.Threads; -084import org.apache.hadoop.io.IOUtils; -085import org.apache.hadoop.io.IntWritable; -086import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -087import org.apache.hadoop.util.StringUtils; -088import org.apache.htrace.TraceInfo; -089 -090import com.google.common.util.concurrent.ThreadFactoryBuilder; +018package org.apache.hadoop.hbase.ipc; +019 +020import java.io.IOException; +021import java.net.BindException; +022import java.net.InetSocketAddress; +023import java.net.ServerSocket; +024import java.net.SocketException; +025import java.net.UnknownHostException; +026import java.nio.channels.CancelledKeyException; +027import java.nio.channels.GatheringByteChannel; +028import java.nio.channels.SelectionKey; +029import java.nio.channels.Selector; +030import java.nio.channels.ServerSocketChannel; +031import java.nio.channels.SocketChannel; +032import java.util.Collections; +033import java.util.Iterator; +034import java.util.List; +035import java.util.Set; +036import java.util.Timer; +037import java.util.TimerTask; +038import
[01/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
Repository: hbase-site Updated Branches: refs/heads/asf-site dd8334d37 -> f55ebeaa5 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/testdevapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html index 8335bdf..e49c7e6 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html @@ -323,138 +323,139 @@ 315 new InetSocketAddress("localhost", 0), conf, scheduler); 316} 317 -318class FailingConnection extends Connection { -319 public FailingConnection(SocketChannel channel, long lastContact) { -320super(channel, lastContact); -321 } -322 -323 @Override -324 public void processRequest(ByteBuff buf) throws IOException, InterruptedException { -325// this will throw exception after the connection header is read, and an RPC is sent -326// from client -327throw new DoNotRetryIOException("Failing for test"); -328 } -329} -330 -331@Override -332protected Connection getConnection(SocketChannel channel, long time) { -333 return new FailingConnection(channel, time); -334} -335 } -336 -337 /** Tests that the connection closing is handled by the client with outstanding RPC calls */ -338 @Test -339 public void testConnectionCloseWithOutstandingRPCs() throws InterruptedException, IOException { -340Configuration conf = new Configuration(CONF); -341RpcServer rpcServer = new TestFailingRpcServer(conf); -342try (AbstractRpcClient? client = createRpcClient(conf)) { -343 rpcServer.start(); -344 BlockingInterface stub = newBlockingStub(client, rpcServer.getListenerAddress()); -345 EchoRequestProto param = EchoRequestProto.newBuilder().setMessage("hello").build(); -346 stub.echo(null, param); -347 fail("RPC should have failed because connection closed"); -348} catch (ServiceException e) { -349 LOG.info("Caught expected exception: " + e.toString()); -350} finally { -351 rpcServer.stop(); -352} -353 } -354 -355 @Test -356 public void testAsyncEcho() throws IOException { -357Configuration conf = HBaseConfiguration.create(); -358RpcServer rpcServer = RpcServerFactory.createRpcServer(null, -359"testRpcServer", Lists.newArrayList(new BlockingServiceAndInterface( -360SERVICE, null)), new InetSocketAddress("localhost", 0), CONF, -361new FifoRpcScheduler(CONF, 1)); -362try (AbstractRpcClient? client = createRpcClient(conf)) { -363 rpcServer.start(); -364 Interface stub = newStub(client, rpcServer.getListenerAddress()); -365 int num = 10; -366 ListHBaseRpcController pcrcList = new ArrayList(); -367 ListBlockingRpcCallbackEchoResponseProto callbackList = new ArrayList(); -368 for (int i = 0; i num; i++) { -369HBaseRpcController pcrc = new HBaseRpcControllerImpl(); -370 BlockingRpcCallbackEchoResponseProto done = new BlockingRpcCallback(); -371stub.echo(pcrc, EchoRequestProto.newBuilder().setMessage("hello-" + i).build(), done); -372pcrcList.add(pcrc); -373callbackList.add(done); -374 } -375 for (int i = 0; i num; i++) { -376HBaseRpcController pcrc = pcrcList.get(i); -377assertFalse(pcrc.failed()); -378assertNull(pcrc.cellScanner()); -379assertEquals("hello-" + i, callbackList.get(i).get().getMessage()); -380 } -381} finally { -382 rpcServer.stop(); -383} -384 } -385 -386 @Test -387 public void testAsyncRemoteError() throws IOException { -388AbstractRpcClient? client = createRpcClient(CONF); -389RpcServer rpcServer = RpcServerFactory.createRpcServer(null, -390"testRpcServer", Lists.newArrayList(new BlockingServiceAndInterface( -391SERVICE, null)), new InetSocketAddress("localhost", 0), CONF, -392new FifoRpcScheduler(CONF, 1)); -393try { -394 rpcServer.start(); -395 Interface stub = newStub(client, rpcServer.getListenerAddress()); -396 BlockingRpcCallbackEmptyResponseProto callback = new BlockingRpcCallback(); -397 HBaseRpcController pcrc = new HBaseRpcControllerImpl(); -398 stub.error(pcrc, EmptyRequestProto.getDefaultInstance(), callback); -399 assertNull(callback.get()); -400 assertTrue(pcrc.failed()); -401 LOG.info("Caught expected exception: " + pcrc.getFailed()); -402 IOException ioe =
[02/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/testdevapidocs/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html b/testdevapidocs/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html index ee3897a..da77885 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html +++ b/testdevapidocs/org/apache/hadoop/hbase/ipc/AbstractTestIPC.TestFailingRpcServer.FailingConnection.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -75,13 +75,13 @@ var activeTableTab = "activeTableTab"; Summary: Nested| -Field| +Field| Constr| Method Detail: -Field| +Field| Constr| Method @@ -100,20 +100,10 @@ var activeTableTab = "activeTableTab"; http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object -org.apache.hadoop.hbase.ipc.RpcServer.Connection - - -org.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection - - org.apache.hadoop.hbase.ipc.AbstractTestIPC.TestFailingRpcServer.FailingConnection - - - - @@ -128,7 +118,7 @@ var activeTableTab = "activeTableTab"; class AbstractTestIPC.TestFailingRpcServer.FailingConnection -extends org.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -141,20 +131,145 @@ extends org.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection Field Summary - - - - -Fields inherited from classorg.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection -channel, responseQueue, socket - - - - - -Fields inherited from classorg.apache.hadoop.hbase.ipc.RpcServer.Connection -addr, attemptingUser, authenticatedWithFallback, authFailedCall, authFailedResponse, authMethod, AUTHORIZATION_FAILED_CALLID, callCleanup, codec, compressionCodec, CONNECTION_HEADER_RESPONSE_CALLID, connectionHeader, connectionHeaderRead, connectionPreambleRead, cryptoAES, hostAddress, remotePort, retryImmediatelySupported, SASL_CALLID, saslCall, saslContextEstablished, saslServer, service, setConnectionHeaderResponseCall, skipInitialSaslHandshake, ugi, useCryptoAesWrap, user, useSasl, useWrap - + +Fields + +Modifier and Type +Field and Description + + +protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddress +addr + + +protected boolean +authenticatedWithFallback + + +protected org.apache.hadoop.hbase.ipc.ServerCall? +authFailedCall + + +protected http://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true; title="class or interface in java.io">ByteArrayOutputStream +authFailedResponse + + +protected org.apache.hadoop.hbase.security.AuthMethod +authMethod + + +protected static int +AUTHORIZATION_FAILED_CALLID + + +protected org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup +callCleanup + + +(package private) http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SocketChannel.html?is-external=true; title="class or interface in java.nio.channels">SocketChannel +channel + + +protected org.apache.hadoop.hbase.codec.Codec +codec + + +protected org.apache.hadoop.io.compress.CompressionCodec +compressionCodec + + +protected static int +CONNECTION_HEADER_RESPONSE_CALLID + + +protected org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader +connectionHeader + + +protected boolean +connectionHeaderRead + + +protected boolean +connectionPreambleRead + + +protected org.apache.hadoop.hbase.io.crypto.aes.CryptoAES +cryptoAES + + +protected http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +hostAddress + + +protected int +remotePort + + +protected http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentLinkedDeque.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentLinkedDequeorg.apache.hadoop.hbase.ipc.SimpleServerCall +responseQueue + + +(package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">Lock +responseWriteLock + +
[41/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/ServerRpcConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/ServerRpcConnection.html b/devapidocs/org/apache/hadoop/hbase/ipc/ServerRpcConnection.html new file mode 100644 index 000..12f89ea --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/ipc/ServerRpcConnection.html @@ -0,0 +1,1181 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +ServerRpcConnection (Apache HBase 2.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":6,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":6,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.ipc +Class ServerRpcConnection + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.ipc.ServerRpcConnection + + + + + + + +All Implemented Interfaces: +http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable, http://docs.oracle.com/javase/8/docs/api/java/lang/AutoCloseable.html?is-external=true; title="class or interface in java.lang">AutoCloseable + + +Direct Known Subclasses: +NettyServerRpcConnection, SimpleServerRpcConnection + + + +abstract class ServerRpcConnection +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable +Reads calls from a connection and queues them for handling. + + + + + + + + + + + +Nested Class Summary + +Nested Classes + +Modifier and Type +Class and Description + + +private static class +ServerRpcConnection.ByteBuffByteInput + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddress +addr + + +private org.apache.hadoop.security.UserGroupInformation +attemptingUser + + +protected boolean +authenticatedWithFallback + + +protected ServerCall? +authFailedCall + + +protected http://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true; title="class or interface in java.io">ByteArrayOutputStream +authFailedResponse + + +protected AuthMethod +authMethod + + +protected static int +AUTHORIZATION_FAILED_CALLID + + +protected RpcServer.CallCleanup +callCleanup + + +protected Codec +codec +Codec the client asked use. + + + +protected org.apache.hadoop.io.compress.CompressionCodec +compressionCodec +Compression codec the client asked us use. + + + +protected static int +CONNECTION_HEADER_RESPONSE_CALLID + + +protected org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader +connectionHeader + + +protected boolean +connectionHeaderRead + + +protected boolean +connectionPreambleRead + + +protected CryptoAES +cryptoAES + + +protected http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +hostAddress + + +protected int +remotePort + + +protected boolean +retryImmediatelySupported + + +protected RpcServer +rpcServer + + +protected static int +SASL_CALLID + + +protected ServerCall? +saslCall + + +protected boolean +saslContextEstablished + + +protected http://docs.oracle.com/javase/8/docs/api/javax/security/sasl/SaslServer.html?is-external=true; title="class or interface in javax.security.sasl">SaslServer +saslServer + + +protected org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService +service + + +protected ServerCall?
[40/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html deleted file mode 100644 index eff1cce..000 --- a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html +++ /dev/null @@ -1,706 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - - - - -SimpleRpcServer.Connection (Apache HBase 2.0.0-SNAPSHOT API) - - - - - -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; -var altColor = "altColor"; -var rowColor = "rowColor"; -var tableTab = "tableTab"; -var activeTableTab = "activeTableTab"; - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -PrevClass -NextClass - - -Frames -NoFrames - - -AllClasses - - - - - - - -Summary: -Nested| -Field| -Constr| -Method - - -Detail: -Field| -Constr| -Method - - - - - - - - -org.apache.hadoop.hbase.ipc -Class SimpleRpcServer.Connection - - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object - - -org.apache.hadoop.hbase.ipc.RpcServer.Connection - - -org.apache.hadoop.hbase.ipc.SimpleRpcServer.Connection - - - - - - - - - -All Implemented Interfaces: -http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable, http://docs.oracle.com/javase/8/docs/api/java/lang/AutoCloseable.html?is-external=true; title="class or interface in java.lang">AutoCloseable - - -Enclosing class: -SimpleRpcServer - - - -public class SimpleRpcServer.Connection -extends RpcServer.Connection -Reads calls from a connection and queues them for handling. - - - - - - - - - - - -Field Summary - -Fields - -Modifier and Type -Field and Description - - -protected http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SocketChannel.html?is-external=true; title="class or interface in java.nio.channels">SocketChannel -channel - - -private ByteBuff -data - - -private http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBuffer -dataLengthBuffer - - -private long -lastContact - - -protected http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentLinkedDeque.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentLinkedDequeSimpleServerCall -responseQueue - - -private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">Lock -responseWriteLock - - -private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true; title="class or interface in java.util.concurrent.atomic">LongAdder -rpcCount - - -protected http://docs.oracle.com/javase/8/docs/api/java/net/Socket.html?is-external=true; title="class or interface in java.net">Socket -socket - - - - - - -Fields inherited from classorg.apache.hadoop.hbase.ipc.RpcServer.Connection -addr, attemptingUser, authenticatedWithFallback, authFailedCall, authFailedResponse, authMethod, AUTHORIZATION_FAILED_CALLID, callCleanup, codec, compressionCodec, CONNECTION_HEADER_RESPONSE_CALLID, connectionHeader, connectionHeaderRead, connectionPreambleRead, cryptoAES, hostAddress, remotePort, retryImmediatelySupported, SASL_CALLID, saslCall, saslContextEstablished, saslServer, service, setConnectionHeaderResponseCall, skipInitialSaslHandshake, ugi, useCryptoAesWrap, user, useSasl, useWrap - - - - - - - - -Constructor Summary - -Constructors - -Constructor and Description - - -Connection(http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SocketChannel.html?is-external=true; title="class or interface in java.nio.channels">SocketChannelchannel, - longlastContact) - - - - - - - - - -Method Summary - -All MethodsInstance MethodsConcrete Methods - -Modifier and Type -Method and Description - - -protected int
[29/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html index 109b5f3..e484176 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.Initializer.html @@ -23,484 +23,310 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import io.netty.bootstrap.ServerBootstrap; -022import io.netty.buffer.ByteBuf; -023import io.netty.buffer.PooledByteBufAllocator; -024import io.netty.buffer.Unpooled; -025import io.netty.channel.Channel; -026import io.netty.channel.ChannelFuture; -027import io.netty.channel.ChannelFutureListener; -028import io.netty.channel.ChannelHandlerContext; -029import io.netty.channel.ChannelInboundHandlerAdapter; -030import io.netty.channel.ChannelInitializer; -031import io.netty.channel.ChannelOption; -032import io.netty.channel.ChannelOutboundHandlerAdapter; -033import io.netty.channel.ChannelPipeline; -034import io.netty.channel.ChannelPromise; -035import io.netty.channel.EventLoopGroup; -036import io.netty.channel.epoll.EpollEventLoopGroup; -037import io.netty.channel.epoll.EpollServerSocketChannel; -038import io.netty.channel.group.ChannelGroup; -039import io.netty.channel.group.DefaultChannelGroup; -040import io.netty.channel.nio.NioEventLoopGroup; -041import io.netty.channel.socket.SocketChannel; -042import io.netty.channel.socket.nio.NioServerSocketChannel; -043import io.netty.handler.codec.ByteToMessageDecoder; -044import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -045import io.netty.util.concurrent.GlobalEventExecutor; -046 -047import java.io.IOException; -048import java.io.InterruptedIOException; -049import java.net.InetAddress; -050import java.net.InetSocketAddress; -051import java.nio.ByteBuffer; -052import java.util.Arrays; -053import java.util.List; -054import java.util.concurrent.CountDownLatch; -055 -056import org.apache.commons.logging.Log; -057import org.apache.commons.logging.LogFactory; -058import org.apache.hadoop.conf.Configuration; -059import org.apache.hadoop.hbase.CellScanner; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -063import org.apache.hadoop.hbase.nio.ByteBuff; -064import org.apache.hadoop.hbase.nio.SingleByteBuff; -065import org.apache.hadoop.hbase.security.AccessDeniedException; -066import org.apache.hadoop.hbase.security.AuthMethod; -067import org.apache.hadoop.hbase.security.HBasePolicyProvider; -068import org.apache.hadoop.hbase.security.SaslStatus; -069import org.apache.hadoop.hbase.security.SaslUtil; -070import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -073import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -074import org.apache.hadoop.hbase.util.Bytes; -075import org.apache.hadoop.hbase.util.JVM; -076import org.apache.hadoop.hbase.util.Pair; -077import org.apache.hadoop.io.IntWritable; -078import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -079import org.apache.htrace.TraceInfo; +018package org.apache.hadoop.hbase.ipc; +019 +020import io.netty.bootstrap.ServerBootstrap; +021import io.netty.buffer.ByteBuf; +022import io.netty.buffer.PooledByteBufAllocator; +023import io.netty.buffer.Unpooled; +024import io.netty.channel.Channel; +025import io.netty.channel.ChannelFuture; +026import io.netty.channel.ChannelFutureListener; +027import io.netty.channel.ChannelHandlerContext; +028import io.netty.channel.ChannelInboundHandlerAdapter; +029import io.netty.channel.ChannelInitializer; +030import io.netty.channel.ChannelOption; +031import io.netty.channel.ChannelOutboundHandlerAdapter; +032import io.netty.channel.ChannelPipeline; +033import io.netty.channel.ChannelPromise; +034import io.netty.channel.EventLoopGroup; +035import io.netty.channel.epoll.EpollEventLoopGroup; +036import io.netty.channel.epoll.EpollServerSocketChannel; +037import io.netty.channel.group.ChannelGroup; +038import io.netty.channel.group.DefaultChannelGroup; +039import io.netty.channel.nio.NioEventLoopGroup; +040import io.netty.channel.socket.SocketChannel; +041import io.netty.channel.socket.nio.NioServerSocketChannel; +042import io.netty.handler.codec.ByteToMessageDecoder; +043import
[33/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/security/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html index e40986a..1c19e45 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html @@ -191,9 +191,9 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) +org.apache.hadoop.hbase.security.SaslStatus org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection org.apache.hadoop.hbase.security.AuthMethod -org.apache.hadoop.hbase.security.SaslStatus http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html index acb7749..8068c8f 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html @@ -198,9 +198,9 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html b/devapidocs/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html index 5e43c3c..8689ac1 100644 --- a/devapidocs/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html +++ b/devapidocs/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.ImplData.html @@ -393,40 +393,40 @@ extends org.jamon.AbstractTemplateProxy.ImplData privateHMaster m_master - + -m_serverManager -privateServerManager m_serverManager +m_deadServers +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetServerName m_deadServers - + -m_serverManager__IsNotDefault -privateboolean m_serverManager__IsNotDefault +m_deadServers__IsNotDefault +privateboolean m_deadServers__IsNotDefault - + -m_format -privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String m_format +m_frags +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer m_frags - + -m_format__IsNotDefault -privateboolean m_format__IsNotDefault +m_frags__IsNotDefault +privateboolean m_frags__IsNotDefault @@ -447,112 +447,112 @@ extends org.jamon.AbstractTemplateProxy.ImplData privateboolean m_catalogJanitorEnabled__IsNotDefault - + -m_servers -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName m_servers +m_format +privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String m_format - + -m_servers__IsNotDefault -privateboolean m_servers__IsNotDefault +m_format__IsNotDefault +privateboolean m_format__IsNotDefault - + -m_filter -privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String
[44/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html deleted file mode 100644 index 9239da0..000 --- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.Connection.html +++ /dev/null @@ -1,1155 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - - - - -RpcServer.Connection (Apache HBase 2.0.0-SNAPSHOT API) - - - - - -var methods = {"i0":10,"i1":6,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":6,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]}; -var altColor = "altColor"; -var rowColor = "rowColor"; -var tableTab = "tableTab"; -var activeTableTab = "activeTableTab"; - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -PrevClass -NextClass - - -Frames -NoFrames - - -AllClasses - - - - - - - -Summary: -Nested| -Field| -Constr| -Method - - -Detail: -Field| -Constr| -Method - - - - - - - - -org.apache.hadoop.hbase.ipc -Class RpcServer.Connection - - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object - - -org.apache.hadoop.hbase.ipc.RpcServer.Connection - - - - - - - -All Implemented Interfaces: -http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable, http://docs.oracle.com/javase/8/docs/api/java/lang/AutoCloseable.html?is-external=true; title="class or interface in java.lang">AutoCloseable - - -Direct Known Subclasses: -NettyRpcServer.NettyConnection, SimpleRpcServer.Connection - - -Enclosing class: -RpcServer - - - -public abstract class RpcServer.Connection -extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object -implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable -Reads calls from a connection and queues them for handling. - - - - - - - - - - - -Field Summary - -Fields - -Modifier and Type -Field and Description - - -protected http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddress -addr - - -org.apache.hadoop.security.UserGroupInformation -attemptingUser - - -protected boolean -authenticatedWithFallback - - -protected ServerCall -authFailedCall - - -protected http://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true; title="class or interface in java.io">ByteArrayOutputStream -authFailedResponse - - -protected AuthMethod -authMethod - - -protected static int -AUTHORIZATION_FAILED_CALLID - - -protected RpcServer.CallCleanup -callCleanup - - -protected Codec -codec -Codec the client asked use. - - - -protected org.apache.hadoop.io.compress.CompressionCodec -compressionCodec -Compression codec the client asked us use. - - - -protected static int -CONNECTION_HEADER_RESPONSE_CALLID - - -protected org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader -connectionHeader - - -protected boolean -connectionHeaderRead - - -protected boolean -connectionPreambleRead - - -protected CryptoAES -cryptoAES - - -protected http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String -hostAddress - - -protected int -remotePort - - -protected boolean -retryImmediatelySupported - - -protected static int -SASL_CALLID - - -protected ServerCall -saslCall - - -protected boolean -saslContextEstablished - - -protected http://docs.oracle.com/javase/8/docs/api/javax/security/sasl/SaslServer.html?is-external=true; title="class or interface in javax.security.sasl">SaslServer -saslServer - - -protected org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService -service - - -protected ServerCall -setConnectionHeaderResponseCall - - -protected boolean -skipInitialSaslHandshake - - -protected org.apache.hadoop.security.UserGroupInformation -ugi - - -private
[16/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerRpcConnection.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerRpcConnection.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerRpcConnection.html new file mode 100644 index 000..50c4422 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerRpcConnection.html @@ -0,0 +1,924 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.ipc; +019 +020import java.io.ByteArrayInputStream; +021import java.io.ByteArrayOutputStream; +022import java.io.Closeable; +023import java.io.DataOutputStream; +024import java.io.IOException; +025import java.net.InetAddress; +026import java.net.InetSocketAddress; +027import java.nio.ByteBuffer; +028import java.nio.channels.Channels; +029import java.nio.channels.ReadableByteChannel; +030import java.security.GeneralSecurityException; +031import java.security.PrivilegedExceptionAction; +032import java.util.Properties; +033 +034import javax.security.sasl.Sasl; +035import javax.security.sasl.SaslException; +036import javax.security.sasl.SaslServer; +037 +038import org.apache.commons.crypto.cipher.CryptoCipherFactory; +039import org.apache.commons.crypto.random.CryptoRandom; +040import org.apache.commons.crypto.random.CryptoRandomFactory; +041import org.apache.hadoop.hbase.CellScanner; +042import org.apache.hadoop.hbase.DoNotRetryIOException; +043import org.apache.hadoop.hbase.client.VersionInfoUtil; +044import org.apache.hadoop.hbase.codec.Codec; +045import org.apache.hadoop.hbase.io.ByteBufferOutputStream; +046import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; +047import org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup; +048import org.apache.hadoop.hbase.nio.ByteBuff; +049import org.apache.hadoop.hbase.nio.SingleByteBuff; +050import org.apache.hadoop.hbase.security.AccessDeniedException; +051import org.apache.hadoop.hbase.security.AuthMethod; +052import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; +053import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; +054import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; +055import org.apache.hadoop.hbase.security.SaslStatus; +056import org.apache.hadoop.hbase.security.SaslUtil; +057import org.apache.hadoop.hbase.security.User; +058import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; +059import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; +060import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; +061import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; +062import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +063import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; +064import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; +065import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; +066import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; +067import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; +068import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; +069import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; +070import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; +071import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation; +072import org.apache.hadoop.hbase.util.Bytes; +073import org.apache.hadoop.io.BytesWritable; +074import org.apache.hadoop.io.Writable; +075import org.apache.hadoop.io.WritableUtils; +076import org.apache.hadoop.io.compress.CompressionCodec; +077import org.apache.hadoop.security.UserGroupInformation; +078import org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod; +079import org.apache.hadoop.security.authorize.AuthorizationException;
[22/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html deleted file mode 100644 index 9e1c66c..000 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.ByteBuffByteInput.html +++ /dev/null @@ -1,1689 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - -Source code - - - - -001/** -002 * Licensed to the Apache Software Foundation (ASF) under one -003 * or more contributor license agreements. See the NOTICE file -004 * distributed with this work for additional information -005 * regarding copyright ownership. The ASF licenses this file -006 * to you under the Apache License, Version 2.0 (the -007 * "License"); you may not use this file except in compliance -008 * with the License. You may obtain a copy of the License at -009 * -010 * http://www.apache.org/licenses/LICENSE-2.0 -011 * -012 * Unless required by applicable law or agreed to in writing, software -013 * distributed under the License is distributed on an "AS IS" BASIS, -014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -015 * See the License for the specific language governing permissions and -016 * limitations under the License. -017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHORIZATION; -022 -023import java.io.ByteArrayInputStream; -024import java.io.ByteArrayOutputStream; -025import java.io.Closeable; -026import java.io.DataOutputStream; -027import java.io.IOException; -028import java.net.InetAddress; -029import java.net.InetSocketAddress; -030import java.nio.ByteBuffer; -031import java.nio.channels.Channels; -032import java.nio.channels.GatheringByteChannel; -033import java.nio.channels.ReadableByteChannel; -034import java.nio.channels.WritableByteChannel; -035import java.security.GeneralSecurityException; -036import java.security.PrivilegedExceptionAction; -037import java.util.ArrayList; -038import java.util.HashMap; -039import java.util.List; -040import java.util.Map; -041import java.util.Properties; -042import java.util.concurrent.atomic.LongAdder; -043 -044import javax.security.sasl.Sasl; -045import javax.security.sasl.SaslException; -046import javax.security.sasl.SaslServer; -047 -048import org.apache.commons.crypto.cipher.CryptoCipherFactory; -049import org.apache.commons.crypto.random.CryptoRandom; -050import org.apache.commons.crypto.random.CryptoRandomFactory; -051import org.apache.commons.logging.Log; -052import org.apache.commons.logging.LogFactory; -053import org.apache.hadoop.conf.Configuration; -054import org.apache.hadoop.hbase.CallQueueTooBigException; -055import org.apache.hadoop.hbase.CellScanner; -056import org.apache.hadoop.hbase.DoNotRetryIOException; -057import org.apache.hadoop.hbase.HBaseInterfaceAudience; -058import org.apache.hadoop.hbase.HConstants; -059import org.apache.hadoop.hbase.Server; -060import org.apache.hadoop.hbase.classification.InterfaceAudience; -061import org.apache.hadoop.hbase.classification.InterfaceStability; -062import org.apache.hadoop.hbase.client.VersionInfoUtil; -063import org.apache.hadoop.hbase.codec.Codec; -064import org.apache.hadoop.hbase.conf.ConfigurationObserver; -065import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -066import org.apache.hadoop.hbase.io.ByteBufferOutputStream; -067import org.apache.hadoop.hbase.io.ByteBufferPool; -068import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; -069import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -070import org.apache.hadoop.hbase.monitoring.TaskMonitor; -071import org.apache.hadoop.hbase.nio.ByteBuff; -072import org.apache.hadoop.hbase.nio.MultiByteBuff; -073import org.apache.hadoop.hbase.nio.SingleByteBuff; -074import org.apache.hadoop.hbase.regionserver.RSRpcServices; -075import org.apache.hadoop.hbase.security.AccessDeniedException; -076import org.apache.hadoop.hbase.security.AuthMethod; -077import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; -078import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; -079import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; -080import org.apache.hadoop.hbase.security.SaslStatus; -081import org.apache.hadoop.hbase.security.SaslUtil; -082import org.apache.hadoop.hbase.security.User; -083import org.apache.hadoop.hbase.security.UserProvider; -084import org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager; -085import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -086import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput;
[14/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html index 7f61b54..35e8890 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.ConnectionManager.html @@ -23,1363 +23,690 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import java.io.IOException; -022import java.io.InputStream; -023import java.net.BindException; -024import java.net.InetAddress; -025import java.net.InetSocketAddress; -026import java.net.ServerSocket; -027import java.net.Socket; -028import java.net.SocketException; -029import java.net.UnknownHostException; -030import java.nio.ByteBuffer; -031import java.nio.channels.CancelledKeyException; -032import java.nio.channels.ClosedChannelException; -033import java.nio.channels.GatheringByteChannel; -034import java.nio.channels.ReadableByteChannel; -035import java.nio.channels.SelectionKey; -036import java.nio.channels.Selector; -037import java.nio.channels.ServerSocketChannel; -038import java.nio.channels.SocketChannel; -039import java.util.ArrayList; -040import java.util.Arrays; -041import java.util.Collections; -042import java.util.Iterator; -043import java.util.List; -044import java.util.Set; -045import java.util.Timer; -046import java.util.TimerTask; -047import java.util.concurrent.ConcurrentHashMap; -048import java.util.concurrent.ConcurrentLinkedDeque; -049import java.util.concurrent.ExecutorService; -050import java.util.concurrent.Executors; -051import java.util.concurrent.LinkedBlockingQueue; -052import java.util.concurrent.atomic.AtomicInteger; -053import java.util.concurrent.atomic.LongAdder; -054import java.util.concurrent.locks.Lock; -055import java.util.concurrent.locks.ReentrantLock; -056 -057import org.apache.hadoop.conf.Configuration; -058import org.apache.hadoop.hbase.CellScanner; -059import org.apache.hadoop.hbase.DoNotRetryIOException; -060import org.apache.hadoop.hbase.HBaseIOException; -061import org.apache.hadoop.hbase.HConstants; -062import org.apache.hadoop.hbase.Server; -063import org.apache.hadoop.hbase.classification.InterfaceAudience; -064import org.apache.hadoop.hbase.classification.InterfaceStability; -065import org.apache.hadoop.hbase.client.VersionInfoUtil; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -068import org.apache.hadoop.hbase.nio.ByteBuff; -069import org.apache.hadoop.hbase.nio.SingleByteBuff; -070import org.apache.hadoop.hbase.security.AccessDeniedException; -071import org.apache.hadoop.hbase.security.AuthMethod; -072import org.apache.hadoop.hbase.security.HBasePolicyProvider; -073import org.apache.hadoop.hbase.security.SaslStatus; -074import org.apache.hadoop.hbase.security.SaslUtil; -075import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -076import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -077import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -078import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -079import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -081import org.apache.hadoop.hbase.util.Bytes; -082import org.apache.hadoop.hbase.util.Pair; -083import org.apache.hadoop.hbase.util.Threads; -084import org.apache.hadoop.io.IOUtils; -085import org.apache.hadoop.io.IntWritable; -086import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -087import org.apache.hadoop.util.StringUtils; -088import org.apache.htrace.TraceInfo; -089 -090import com.google.common.util.concurrent.ThreadFactoryBuilder; +018package org.apache.hadoop.hbase.ipc; +019 +020import java.io.IOException; +021import java.net.BindException; +022import java.net.InetSocketAddress; +023import java.net.ServerSocket; +024import java.net.SocketException; +025import java.net.UnknownHostException; +026import java.nio.channels.CancelledKeyException; +027import java.nio.channels.GatheringByteChannel; +028import java.nio.channels.SelectionKey; +029import java.nio.channels.Selector; +030import java.nio.channels.ServerSocketChannel; +031import java.nio.channels.SocketChannel; +032import java.util.Collections; +033import java.util.Iterator; +034import java.util.List; +035import java.util.Set; +036import java.util.Timer; +037import java.util.TimerTask;
[09/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.html new file mode 100644 index 000..e79bf23 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServerResponder.html @@ -0,0 +1,388 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.ipc; +019 +020import java.io.IOException; +021import java.nio.channels.CancelledKeyException; +022import java.nio.channels.ClosedChannelException; +023import java.nio.channels.SelectionKey; +024import java.nio.channels.Selector; +025import java.util.ArrayList; +026import java.util.Collections; +027import java.util.Iterator; +028import java.util.Set; +029import java.util.concurrent.ConcurrentHashMap; +030 +031import org.apache.hadoop.hbase.HBaseIOException; +032import org.apache.hadoop.hbase.classification.InterfaceAudience; +033import org.apache.hadoop.hbase.util.Threads; +034import org.apache.hadoop.util.StringUtils; +035 +036/** +037 * Sends responses of RPC back to clients. +038 */ +039@InterfaceAudience.Private +040class SimpleRpcServerResponder extends Thread { +041 /** */ +042 private final SimpleRpcServer simpleRpcServer; +043 private final Selector writeSelector; +044 private final SetSimpleServerRpcConnection writingCons = +045 Collections.newSetFromMap(new ConcurrentHashMapSimpleServerRpcConnection, Boolean()); +046 +047 SimpleRpcServerResponder(SimpleRpcServer simpleRpcServer) throws IOException { +048this.simpleRpcServer = simpleRpcServer; +049 this.setName("RpcServer.responder"); +050this.setDaemon(true); +051 this.setUncaughtExceptionHandler(Threads.LOGGING_EXCEPTION_HANDLER); +052writeSelector = Selector.open(); // create a selector +053 } +054 +055 @Override +056 public void run() { +057SimpleRpcServer.LOG.debug(getName() + ": starting"); +058try { +059 doRunLoop(); +060} finally { +061 SimpleRpcServer.LOG.info(getName() + ": stopping"); +062 try { +063writeSelector.close(); +064 } catch (IOException ioe) { +065 SimpleRpcServer.LOG.error(getName() + ": couldn't close write selector", ioe); +066 } +067} +068 } +069 +070 /** +071 * Take the list of the connections that want to write, and register them in the selector. +072 */ +073 private void registerWrites() { +074 IteratorSimpleServerRpcConnection it = writingCons.iterator(); +075while (it.hasNext()) { +076 SimpleServerRpcConnection c = it.next(); +077 it.remove(); +078 SelectionKey sk = c.channel.keyFor(writeSelector); +079 try { +080if (sk == null) { +081 try { +082 c.channel.register(writeSelector, SelectionKey.OP_WRITE, c); +083 } catch (ClosedChannelException e) { +084// ignore: the client went away. +085if (SimpleRpcServer.LOG.isTraceEnabled()) SimpleRpcServer.LOG.trace("ignored", e); +086 } +087} else { +088 sk.interestOps(SelectionKey.OP_WRITE); +089} +090 } catch (CancelledKeyException e) { +091// ignore: the client went away. +092if (SimpleRpcServer.LOG.isTraceEnabled()) SimpleRpcServer.LOG.trace("ignored", e); +093 } +094} +095 } +096 +097 /** +098 * Add a connection to the list that want to write, +099 */ +100 public void registerForWrite(SimpleServerRpcConnection c) { +101if (writingCons.add(c)) { +102 writeSelector.wakeup(); +103} +104 } +105 +106 private void doRunLoop() { +107long lastPurgeTime = 0; // last check for old calls. +108while (this.simpleRpcServer.running) { +109 try { +110registerWrites(); +111int keyCt = writeSelector.select(this.simpleRpcServer.purgeTimeout); +112
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html index 6ad1e1b..f2ec99b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/ServerCall.html @@ -38,501 +38,500 @@ 030import org.apache.hadoop.hbase.io.ByteBufferListOutputStream; 031import org.apache.hadoop.hbase.io.ByteBufferPool; 032import org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup; -033import org.apache.hadoop.hbase.ipc.RpcServer.Connection; -034import org.apache.hadoop.hbase.security.User; -035import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -036import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream; -037import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -038import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -039import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -040import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -041import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta; -042import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse; -043import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -044import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader; -045import org.apache.hadoop.hbase.util.ByteBufferUtils; -046import org.apache.hadoop.hbase.util.Bytes; -047import org.apache.hadoop.util.StringUtils; -048import org.apache.htrace.TraceInfo; -049 -050/** -051 * Datastructure that holds all necessary to a method invocation and then afterward, carries -052 * the result. -053 */ -054@InterfaceAudience.Private -055abstract class ServerCall implements RpcCall { -056 -057 protected final int id; // the client's call id -058 protected final BlockingService service; -059 protected final MethodDescriptor md; -060 protected final RequestHeader header; -061 protected Message param; // the parameter passed -062 // Optional cell data passed outside of protobufs. -063 protected final CellScanner cellScanner; -064 protected final Connection connection; // connection to client -065 protected final long receiveTime; // the time received when response is null -066 // the time served when response is not null -067 protected final int timeout; -068 protected long startTime; -069 protected final long deadline;// the deadline to handle this call, if exceed we can drop it. -070 -071 protected final ByteBufferPool reservoir; -072 -073 protected final CellBlockBuilder cellBlockBuilder; -074 -075 /** -076 * Chain of buffers to send as response. -077 */ -078 protected BufferChain response; -079 -080 protected final long size; // size of current call -081 protected boolean isError; -082 protected final TraceInfo tinfo; -083 protected ByteBufferListOutputStream cellBlockStream = null; -084 protected CallCleanup reqCleanup = null; -085 -086 protected User user; -087 protected final InetAddress remoteAddress; -088 protected RpcCallback rpcCallback; -089 -090 private long responseCellSize = 0; -091 private long responseBlockSize = 0; -092 // cumulative size of serialized exceptions -093 private long exceptionSize = 0; -094 private final boolean retryImmediatelySupported; -095 -096 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH", -097 justification="Can't figure why this complaint is happening... see below") -098 ServerCall(int id, BlockingService service, MethodDescriptor md, RequestHeader header, -099 Message param, CellScanner cellScanner, Connection connection, long size, TraceInfo tinfo, -100 InetAddress remoteAddress, long receiveTime, int timeout, ByteBufferPool reservoir, -101 CellBlockBuilder cellBlockBuilder, CallCleanup reqCleanup) { -102this.id = id; -103this.service = service; -104this.md = md; -105this.header = header; -106this.param = param; -107this.cellScanner = cellScanner; -108this.connection = connection; -109this.receiveTime = receiveTime; -110this.response = null; -111this.isError = false; -112this.size = size; -113this.tinfo = tinfo; -114this.user = connection == null ? null : connection.user; // FindBugs: NP_NULL_ON_SOME_PATH -115this.remoteAddress = remoteAddress; -116this.retryImmediatelySupported = -117connection == null ? false : connection.retryImmediatelySupported; -118this.timeout = timeout; -119this.deadline =
[24/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.html new file mode 100644 index 000..4c7720b --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.html @@ -0,0 +1,278 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.ipc; +019 +020import io.netty.buffer.ByteBuf; +021import io.netty.channel.Channel; +022import io.netty.channel.ChannelFutureListener; +023 +024import java.io.IOException; +025import java.net.InetAddress; +026import java.net.InetSocketAddress; +027import java.nio.ByteBuffer; +028import java.util.Arrays; +029 +030import org.apache.hadoop.hbase.CellScanner; +031import org.apache.hadoop.hbase.HConstants; +032import org.apache.hadoop.hbase.classification.InterfaceAudience; +033import org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup; +034import org.apache.hadoop.hbase.nio.ByteBuff; +035import org.apache.hadoop.hbase.nio.SingleByteBuff; +036import org.apache.hadoop.hbase.security.AccessDeniedException; +037import org.apache.hadoop.hbase.security.AuthMethod; +038import org.apache.hadoop.hbase.security.SaslStatus; +039import org.apache.hadoop.hbase.security.SaslUtil; +040import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; +041import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; +042import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +043import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; +044import org.apache.hadoop.hbase.util.Bytes; +045import org.apache.hadoop.io.IntWritable; +046import org.apache.htrace.TraceInfo; +047 +048/** +049 * RpcConnection implementation for netty rpc server. +050 */ +051@InterfaceAudience.Private +052class NettyServerRpcConnection extends ServerRpcConnection { +053 +054 final Channel channel; +055 +056 NettyServerRpcConnection(NettyRpcServer rpcServer, Channel channel) { +057super(rpcServer); +058this.channel = channel; +059InetSocketAddress inetSocketAddress = ((InetSocketAddress) channel.remoteAddress()); +060this.addr = inetSocketAddress.getAddress(); +061if (addr == null) { +062 this.hostAddress = "*Unknown*"; +063} else { +064 this.hostAddress = inetSocketAddress.getAddress().getHostAddress(); +065} +066this.remotePort = inetSocketAddress.getPort(); +067this.saslCall = new NettyServerCall(SASL_CALLID, null, null, null, null, null, this, 0, null, +068null, System.currentTimeMillis(), 0, rpcServer.reservoir, rpcServer.cellBlockBuilder, null); +069this.setConnectionHeaderResponseCall = new NettyServerCall(CONNECTION_HEADER_RESPONSE_CALLID, +070null, null, null, null, null, this, 0, null, null, System.currentTimeMillis(), 0, +071rpcServer.reservoir, rpcServer.cellBlockBuilder, null); +072this.authFailedCall = new NettyServerCall(AUTHORIZATION_FAILED_CALLID, null, null, null, null, +073null, this, 0, null, null, System.currentTimeMillis(), 0, rpcServer.reservoir, +074rpcServer.cellBlockBuilder, null); +075 } +076 +077 void readPreamble(ByteBuf buffer) throws IOException { +078byte[] rpcHead = { buffer.readByte(), buffer.readByte(), buffer.readByte(), buffer.readByte() }; +079if (!Arrays.equals(HConstants.RPC_HEADER, rpcHead)) { +080 doBadPreambleHandling("Expected HEADER=" + Bytes.toStringBinary(HConstants.RPC_HEADER) + +081 " but received HEADER=" + Bytes.toStringBinary(rpcHead) + " from " + toString()); +082 return; +083} +084// Now read the next two bytes, the version and the auth to use. +085int version = buffer.readByte(); +086byte authbyte = buffer.readByte(); +087this.authMethod =
[43/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html index 307b468..fb6e673 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/RpcServer.html @@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.LimitedPrivate(value={"Coprocesssor","Phoenix"}) @InterfaceStability.Evolving -public abstract class RpcServer +public abstract class RpcServer extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements RpcServerInterface, ConfigurationObserver An RPC server that hosts protobuf described Services. @@ -149,19 +149,9 @@ implements -protected static class -RpcServer.ByteBuffByteInput - - protected static interface RpcServer.CallCleanup - -class -RpcServer.Connection -Reads calls from a connection and queues them for handling. - - @@ -633,7 +623,7 @@ implements protected void setupResponse(http://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true; title="class or interface in java.io">ByteArrayOutputStreamresponse, - ServerCallcall, + ServerCall?call, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablet, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringerror) Setup response for the RPC Call. @@ -674,7 +664,7 @@ implements LOG -public static finalorg.apache.commons.logging.Log LOG +public static finalorg.apache.commons.logging.Log LOG @@ -683,7 +673,7 @@ implements CALL_QUEUE_TOO_BIG_EXCEPTION -protected static finalCallQueueTooBigException CALL_QUEUE_TOO_BIG_EXCEPTION +protected static finalCallQueueTooBigException CALL_QUEUE_TOO_BIG_EXCEPTION @@ -692,7 +682,7 @@ implements authorize -private finalboolean authorize +private finalboolean authorize @@ -701,7 +691,7 @@ implements isSecurityEnabled -protectedboolean isSecurityEnabled +protectedboolean isSecurityEnabled @@ -710,7 +700,7 @@ implements CURRENT_VERSION -public static finalbyte CURRENT_VERSION +public static finalbyte CURRENT_VERSION See Also: Constant Field Values @@ -723,7 +713,7 @@ implements FALLBACK_TO_INSECURE_CLIENT_AUTH -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String FALLBACK_TO_INSECURE_CLIENT_AUTH +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String FALLBACK_TO_INSECURE_CLIENT_AUTH Whether we allow a fallback to SIMPLE auth for insecure clients when security is enabled. See Also: @@ -737,7 +727,7 @@ implements DEFAULT_MAX_CALLQUEUE_LENGTH_PER_HANDLER -protected static finalint DEFAULT_MAX_CALLQUEUE_LENGTH_PER_HANDLER +protected static finalint DEFAULT_MAX_CALLQUEUE_LENGTH_PER_HANDLER How many calls/handler are allowed in the queue. See Also: @@ -751,7 +741,7 @@ implements cellBlockBuilder -protected finalCellBlockBuilder cellBlockBuilder +protected finalCellBlockBuilder cellBlockBuilder @@ -760,7 +750,7 @@ implements AUTH_FAILED_FOR -protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String AUTH_FAILED_FOR +protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String AUTH_FAILED_FOR See Also: Constant Field Values @@ -773,7 +763,7 @@ implements AUTH_SUCCESSFUL_FOR -protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String AUTH_SUCCESSFUL_FOR +protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String AUTH_SUCCESSFUL_FOR See Also: Constant Field Values @@ -786,7 +776,7 @@ implements AUDITLOG -protected static finalorg.apache.commons.logging.Log AUDITLOG +protected static finalorg.apache.commons.logging.Log AUDITLOG @@ -795,7 +785,7 @@ implements secretManager -protectedorg.apache.hadoop.security.token.SecretManagerorg.apache.hadoop.security.token.TokenIdentifier secretManager +protectedorg.apache.hadoop.security.token.SecretManagerorg.apache.hadoop.security.token.TokenIdentifier secretManager @@ -804,7 +794,7 @@ implements authManager
[45/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.html b/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.html new file mode 100644 index 000..1d865ec --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/ipc/NettyServerRpcConnection.html @@ -0,0 +1,505 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +NettyServerRpcConnection (Apache HBase 2.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.ipc +Class NettyServerRpcConnection + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.ipc.ServerRpcConnection + + +org.apache.hadoop.hbase.ipc.NettyServerRpcConnection + + + + + + + + + +All Implemented Interfaces: +http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable, http://docs.oracle.com/javase/8/docs/api/java/lang/AutoCloseable.html?is-external=true; title="class or interface in java.lang">AutoCloseable + + + +@InterfaceAudience.Private +class NettyServerRpcConnection +extends ServerRpcConnection +RpcConnection implementation for netty rpc server. + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +(package private) io.netty.channel.Channel +channel + + + + + + +Fields inherited from classorg.apache.hadoop.hbase.ipc.ServerRpcConnection +addr, authenticatedWithFallback, authFailedCall, authFailedResponse, authMethod, AUTHORIZATION_FAILED_CALLID, callCleanup, codec, compressionCodec, CONNECTION_HEADER_RESPONSE_CALLID, connectionHeader, connectionHeaderRead, connectionPreambleRead, cryptoAES, hostAddress, remotePort, retryImmediatelySuppor ted, rpcServer, SASL_CALLID, saslCall, saslContextEstablished, saslServer, service, setConnectionHeaderResponseCall, skipInitialSaslHandshake, ugi, useCryptoAesWrap, user, useSasl, useWrap + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +NettyServerRpcConnection(NettyRpcServerrpcServer, + io.netty.channel.Channelchannel) + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +void +close() + + +NettyServerCall +createCall(intid, + org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingServiceservice, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptormd, + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeaderheader, + org.apache.hadoop.hbase.shaded.com.google.protobuf.Messageparam, + CellScannercellScanner, + longsize, + org.apache.htrace.TraceInfotinfo, + http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddressremoteAddress, + inttimeout, + RpcServer.CallCleanupreqCleanup) + + +private void +doBadPreambleHandling(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringmsg) + + +private void +doBadPreambleHandling(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringmsg, + http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exceptione) + + +boolean +isConnectionOpen() + + +(package private) void +process(io.netty.buffer.ByteBufbuf) + + +(package private) void +process(ByteBuffbuf) + + +(package
[25/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html index 109b5f3..e484176 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.html @@ -23,484 +23,310 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import io.netty.bootstrap.ServerBootstrap; -022import io.netty.buffer.ByteBuf; -023import io.netty.buffer.PooledByteBufAllocator; -024import io.netty.buffer.Unpooled; -025import io.netty.channel.Channel; -026import io.netty.channel.ChannelFuture; -027import io.netty.channel.ChannelFutureListener; -028import io.netty.channel.ChannelHandlerContext; -029import io.netty.channel.ChannelInboundHandlerAdapter; -030import io.netty.channel.ChannelInitializer; -031import io.netty.channel.ChannelOption; -032import io.netty.channel.ChannelOutboundHandlerAdapter; -033import io.netty.channel.ChannelPipeline; -034import io.netty.channel.ChannelPromise; -035import io.netty.channel.EventLoopGroup; -036import io.netty.channel.epoll.EpollEventLoopGroup; -037import io.netty.channel.epoll.EpollServerSocketChannel; -038import io.netty.channel.group.ChannelGroup; -039import io.netty.channel.group.DefaultChannelGroup; -040import io.netty.channel.nio.NioEventLoopGroup; -041import io.netty.channel.socket.SocketChannel; -042import io.netty.channel.socket.nio.NioServerSocketChannel; -043import io.netty.handler.codec.ByteToMessageDecoder; -044import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -045import io.netty.util.concurrent.GlobalEventExecutor; -046 -047import java.io.IOException; -048import java.io.InterruptedIOException; -049import java.net.InetAddress; -050import java.net.InetSocketAddress; -051import java.nio.ByteBuffer; -052import java.util.Arrays; -053import java.util.List; -054import java.util.concurrent.CountDownLatch; -055 -056import org.apache.commons.logging.Log; -057import org.apache.commons.logging.LogFactory; -058import org.apache.hadoop.conf.Configuration; -059import org.apache.hadoop.hbase.CellScanner; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -063import org.apache.hadoop.hbase.nio.ByteBuff; -064import org.apache.hadoop.hbase.nio.SingleByteBuff; -065import org.apache.hadoop.hbase.security.AccessDeniedException; -066import org.apache.hadoop.hbase.security.AuthMethod; -067import org.apache.hadoop.hbase.security.HBasePolicyProvider; -068import org.apache.hadoop.hbase.security.SaslStatus; -069import org.apache.hadoop.hbase.security.SaslUtil; -070import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -073import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -074import org.apache.hadoop.hbase.util.Bytes; -075import org.apache.hadoop.hbase.util.JVM; -076import org.apache.hadoop.hbase.util.Pair; -077import org.apache.hadoop.io.IntWritable; -078import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -079import org.apache.htrace.TraceInfo; +018package org.apache.hadoop.hbase.ipc; +019 +020import io.netty.bootstrap.ServerBootstrap; +021import io.netty.buffer.ByteBuf; +022import io.netty.buffer.PooledByteBufAllocator; +023import io.netty.buffer.Unpooled; +024import io.netty.channel.Channel; +025import io.netty.channel.ChannelFuture; +026import io.netty.channel.ChannelFutureListener; +027import io.netty.channel.ChannelHandlerContext; +028import io.netty.channel.ChannelInboundHandlerAdapter; +029import io.netty.channel.ChannelInitializer; +030import io.netty.channel.ChannelOption; +031import io.netty.channel.ChannelOutboundHandlerAdapter; +032import io.netty.channel.ChannelPipeline; +033import io.netty.channel.ChannelPromise; +034import io.netty.channel.EventLoopGroup; +035import io.netty.channel.epoll.EpollEventLoopGroup; +036import io.netty.channel.epoll.EpollServerSocketChannel; +037import io.netty.channel.group.ChannelGroup; +038import io.netty.channel.group.DefaultChannelGroup; +039import io.netty.channel.nio.NioEventLoopGroup; +040import io.netty.channel.socket.SocketChannel; +041import io.netty.channel.socket.nio.NioServerSocketChannel; +042import io.netty.handler.codec.ByteToMessageDecoder; +043import io.netty.handler.codec.LengthFieldBasedFrameDecoder; +044import
[15/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html deleted file mode 100644 index 7f61b54..000 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/SimpleRpcServer.Connection.html +++ /dev/null @@ -1,1446 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - -Source code - - - - -001/** -002 * Licensed to the Apache Software Foundation (ASF) under one -003 * or more contributor license agreements. See the NOTICE file -004 * distributed with this work for additional information -005 * regarding copyright ownership. The ASF licenses this file -006 * to you under the Apache License, Version 2.0 (the -007 * "License"); you may not use this file except in compliance -008 * with the License. You may obtain a copy of the License at -009 * -010 * http://www.apache.org/licenses/LICENSE-2.0 -011 * -012 * Unless required by applicable law or agreed to in writing, software -013 * distributed under the License is distributed on an "AS IS" BASIS, -014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -015 * See the License for the specific language governing permissions and -016 * limitations under the License. -017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import java.io.IOException; -022import java.io.InputStream; -023import java.net.BindException; -024import java.net.InetAddress; -025import java.net.InetSocketAddress; -026import java.net.ServerSocket; -027import java.net.Socket; -028import java.net.SocketException; -029import java.net.UnknownHostException; -030import java.nio.ByteBuffer; -031import java.nio.channels.CancelledKeyException; -032import java.nio.channels.ClosedChannelException; -033import java.nio.channels.GatheringByteChannel; -034import java.nio.channels.ReadableByteChannel; -035import java.nio.channels.SelectionKey; -036import java.nio.channels.Selector; -037import java.nio.channels.ServerSocketChannel; -038import java.nio.channels.SocketChannel; -039import java.util.ArrayList; -040import java.util.Arrays; -041import java.util.Collections; -042import java.util.Iterator; -043import java.util.List; -044import java.util.Set; -045import java.util.Timer; -046import java.util.TimerTask; -047import java.util.concurrent.ConcurrentHashMap; -048import java.util.concurrent.ConcurrentLinkedDeque; -049import java.util.concurrent.ExecutorService; -050import java.util.concurrent.Executors; -051import java.util.concurrent.LinkedBlockingQueue; -052import java.util.concurrent.atomic.AtomicInteger; -053import java.util.concurrent.atomic.LongAdder; -054import java.util.concurrent.locks.Lock; -055import java.util.concurrent.locks.ReentrantLock; -056 -057import org.apache.hadoop.conf.Configuration; -058import org.apache.hadoop.hbase.CellScanner; -059import org.apache.hadoop.hbase.DoNotRetryIOException; -060import org.apache.hadoop.hbase.HBaseIOException; -061import org.apache.hadoop.hbase.HConstants; -062import org.apache.hadoop.hbase.Server; -063import org.apache.hadoop.hbase.classification.InterfaceAudience; -064import org.apache.hadoop.hbase.classification.InterfaceStability; -065import org.apache.hadoop.hbase.client.VersionInfoUtil; -066import org.apache.hadoop.hbase.exceptions.RequestTooBigException; -067import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -068import org.apache.hadoop.hbase.nio.ByteBuff; -069import org.apache.hadoop.hbase.nio.SingleByteBuff; -070import org.apache.hadoop.hbase.security.AccessDeniedException; -071import org.apache.hadoop.hbase.security.AuthMethod; -072import org.apache.hadoop.hbase.security.HBasePolicyProvider; -073import org.apache.hadoop.hbase.security.SaslStatus; -074import org.apache.hadoop.hbase.security.SaslUtil; -075import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -076import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; -077import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -078import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -079import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -081import org.apache.hadoop.hbase.util.Bytes; -082import org.apache.hadoop.hbase.util.Pair; -083import org.apache.hadoop.hbase.util.Threads; -084import org.apache.hadoop.io.IOUtils; -085import org.apache.hadoop.io.IntWritable; -086import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -087import org.apache.hadoop.util.StringUtils; -088import org.apache.htrace.TraceInfo; -089 -090import
[32/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmplImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmplImpl.html b/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmplImpl.html index 643f553..6f51bbb 100644 --- a/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmplImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmplImpl.html @@ -238,40 +238,40 @@ implements HRegionServer regionServer - + -bcv -private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String bcv +format +private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String format - + -filter -private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String filter +bcn +private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String bcn - + -bcn -private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String bcn +bcv +private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String bcv - + -format -private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String format +filter +private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String filter http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/util/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html index f4cfe34..1af1aa6 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html @@ -519,14 +519,14 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) +org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) +org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE org.apache.hadoop.hbase.util.PrettyPrinter.Unit -org.apache.hadoop.hbase.util.PoolMap.PoolType org.apache.hadoop.hbase.util.ChecksumType +org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType -org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) +org.apache.hadoop.hbase.util.PoolMap.PoolType org.apache.hadoop.hbase.util.Order -org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE -org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/wal/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/wal/package-tree.html b/devapidocs/org/apache/hadoop/hbase/wal/package-tree.html index cb43481..5e6897f 100644 --- a/devapidocs/org/apache/hadoop/hbase/wal/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/wal/package-tree.html @@ -166,8 +166,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.wal.WALFactory.Providers org.apache.hadoop.hbase.wal.RegionGroupingProvider.Strategies +org.apache.hadoop.hbase.wal.WALFactory.Providers
[51/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
Published site at 82d554e3783372cc6b05489452c815b57c06f6cd. Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/f55ebeaa Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/f55ebeaa Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/f55ebeaa Branch: refs/heads/asf-site Commit: f55ebeaa51e6657c89cf6688547512ad24658ea6 Parents: dd8334d Author: jenkinsAuthored: Mon May 15 14:58:34 2017 + Committer: jenkins Committed: Mon May 15 14:58:34 2017 + -- acid-semantics.html |4 +- apache_hbase_reference_guide.pdf|4 +- apache_hbase_reference_guide.pdfmarks |4 +- bulk-loads.html |4 +- checkstyle-aggregate.html | 6190 +- checkstyle.rss | 72 +- coc.html|4 +- cygwin.html |4 +- dependencies.html |4 +- dependency-convergence.html |4 +- dependency-info.html|4 +- dependency-management.html |4 +- devapidocs/allclasses-frame.html|6 +- devapidocs/allclasses-noframe.html |6 +- devapidocs/constant-values.html | 70 +- devapidocs/index-all.html | 318 +- .../hadoop/hbase/backup/package-tree.html |4 +- .../hadoop/hbase/class-use/CellScanner.html | 29 +- .../class-use/InterfaceAudience.Private.html| 26 +- .../hbase/classification/package-tree.html |6 +- .../hadoop/hbase/client/package-tree.html | 24 +- .../hadoop/hbase/codec/class-use/Codec.html |2 +- .../hadoop/hbase/filter/package-tree.html | 10 +- .../hbase/io/class-use/ByteBufferPool.html | 14 +- .../io/crypto/aes/class-use/CryptoAES.html |2 +- .../hadoop/hbase/io/hfile/package-tree.html |6 +- .../org/apache/hadoop/hbase/ipc/CallRunner.html |4 +- .../ipc/NettyRpcServer.CallWriteListener.html |8 +- .../NettyRpcServer.ConnectionHeaderHandler.html | 10 +- .../hbase/ipc/NettyRpcServer.Initializer.html |8 +- .../ipc/NettyRpcServer.MessageDecoder.html | 22 +- .../ipc/NettyRpcServer.MessageEncoder.html | 10 +- .../ipc/NettyRpcServer.NettyConnection.html | 507 -- .../apache/hadoop/hbase/ipc/NettyRpcServer.html | 39 +- .../hadoop/hbase/ipc/NettyServerCall.html | 45 +- .../hbase/ipc/NettyServerRpcConnection.html | 505 ++ .../hadoop/hbase/ipc/PriorityFunction.html |4 +- .../RpcServer.BlockingServiceAndInterface.html | 16 +- .../hbase/ipc/RpcServer.ByteBuffByteInput.html | 460 -- .../hadoop/hbase/ipc/RpcServer.CallCleanup.html | 12 +- .../hadoop/hbase/ipc/RpcServer.Connection.html | 1155 .../org/apache/hadoop/hbase/ipc/RpcServer.html | 184 +- .../hadoop/hbase/ipc/RpcServerFactory.html |4 +- .../org/apache/hadoop/hbase/ipc/ServerCall.html | 152 +- .../hbase/ipc/ServerNotRunningYetException.html |4 +- .../ServerRpcConnection.ByteBuffByteInput.html | 460 ++ .../hadoop/hbase/ipc/ServerRpcConnection.html | 1181 .../hadoop/hbase/ipc/ServerRpcController.html |4 +- .../hbase/ipc/SimpleRpcServer.Connection.html | 706 -- .../ipc/SimpleRpcServer.ConnectionManager.html | 62 +- .../ipc/SimpleRpcServer.Listener.Reader.html| 24 +- .../hbase/ipc/SimpleRpcServer.Listener.html | 30 +- .../hbase/ipc/SimpleRpcServer.Responder.html| 530 -- .../hadoop/hbase/ipc/SimpleRpcServer.html | 72 +- .../hbase/ipc/SimpleRpcServerResponder.html | 536 ++ .../hadoop/hbase/ipc/SimpleServerCall.html | 44 +- .../hbase/ipc/SimpleServerRpcConnection.html| 716 ++ .../hbase/ipc/StoppedRpcClientException.html|4 +- .../hbase/ipc/class-use/CellBlockBuilder.html | 14 +- .../ipc/class-use/FatalConnectionException.html |6 +- .../NettyRpcServer.NettyConnection.html | 197 - .../hbase/ipc/class-use/NettyRpcServer.html | 43 +- .../hbase/ipc/class-use/NettyServerCall.html| 23 + .../ipc/class-use/NettyServerRpcConnection.html | 209 + .../hadoop/hbase/ipc/class-use/RpcCall.html |2 +- .../hbase/ipc/class-use/RpcCallContext.html |2 +- .../class-use/RpcServer.ByteBuffByteInput.html | 125 - .../ipc/class-use/RpcServer.CallCleanup.html| 31 +- .../ipc/class-use/RpcServer.Connection.html | 334 - .../hadoop/hbase/ipc/class-use/RpcServer.html |9 +- .../hadoop/hbase/ipc/class-use/ServerCall.html | 51 +- .../ServerRpcConnection.ByteBuffByteInput.html | 125 + .../ipc/class-use/ServerRpcConnection.html | 204 +
[05/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.html index 23fce63..66a9a30 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/regionserver/RSStatusTmpl.html @@ -33,10 +33,10 @@ 025 requiredArguments = { 026@org.jamon.annotations.Argument(name = "regionServer", type = "HRegionServer")}, 027 optionalArguments = { -028@org.jamon.annotations.Argument(name = "bcv", type = "String"), -029@org.jamon.annotations.Argument(name = "filter", type = "String"), -030@org.jamon.annotations.Argument(name = "bcn", type = "String"), -031@org.jamon.annotations.Argument(name = "format", type = "String")}) +028@org.jamon.annotations.Argument(name = "format", type = "String"), +029@org.jamon.annotations.Argument(name = "bcn", type = "String"), +030@org.jamon.annotations.Argument(name = "bcv", type = "String"), +031@org.jamon.annotations.Argument(name = "filter", type = "String")}) 032public class RSStatusTmpl 033 extends org.jamon.AbstractTemplateProxy 034{ @@ -77,74 +77,74 @@ 069 return m_regionServer; 070} 071private HRegionServer m_regionServer; -072// 24, 1 -073public void setBcv(String bcv) +072// 22, 1 +073public void setFormat(String format) 074{ -075 // 24, 1 -076 m_bcv = bcv; -077 m_bcv__IsNotDefault = true; +075 // 22, 1 +076 m_format = format; +077 m_format__IsNotDefault = true; 078} -079public String getBcv() +079public String getFormat() 080{ -081 return m_bcv; +081 return m_format; 082} -083private String m_bcv; -084public boolean getBcv__IsNotDefault() +083private String m_format; +084public boolean getFormat__IsNotDefault() 085{ -086 return m_bcv__IsNotDefault; +086 return m_format__IsNotDefault; 087} -088private boolean m_bcv__IsNotDefault; -089// 21, 1 -090public void setFilter(String filter) +088private boolean m_format__IsNotDefault; +089// 23, 1 +090public void setBcn(String bcn) 091{ -092 // 21, 1 -093 m_filter = filter; -094 m_filter__IsNotDefault = true; +092 // 23, 1 +093 m_bcn = bcn; +094 m_bcn__IsNotDefault = true; 095} -096public String getFilter() +096public String getBcn() 097{ -098 return m_filter; +098 return m_bcn; 099} -100private String m_filter; -101public boolean getFilter__IsNotDefault() +100private String m_bcn; +101public boolean getBcn__IsNotDefault() 102{ -103 return m_filter__IsNotDefault; +103 return m_bcn__IsNotDefault; 104} -105private boolean m_filter__IsNotDefault; -106// 23, 1 -107public void setBcn(String bcn) +105private boolean m_bcn__IsNotDefault; +106// 24, 1 +107public void setBcv(String bcv) 108{ -109 // 23, 1 -110 m_bcn = bcn; -111 m_bcn__IsNotDefault = true; +109 // 24, 1 +110 m_bcv = bcv; +111 m_bcv__IsNotDefault = true; 112} -113public String getBcn() +113public String getBcv() 114{ -115 return m_bcn; +115 return m_bcv; 116} -117private String m_bcn; -118public boolean getBcn__IsNotDefault() +117private String m_bcv; +118public boolean getBcv__IsNotDefault() 119{ -120 return m_bcn__IsNotDefault; +120 return m_bcv__IsNotDefault; 121} -122private boolean m_bcn__IsNotDefault; -123// 22, 1 -124public void setFormat(String format) +122private boolean m_bcv__IsNotDefault; +123// 21, 1 +124public void setFilter(String filter) 125{ -126 // 22, 1 -127 m_format = format; -128 m_format__IsNotDefault = true; +126 // 21, 1 +127 m_filter = filter; +128 m_filter__IsNotDefault = true; 129} -130public String getFormat() +130public String getFilter() 131{ -132 return m_format; +132 return m_filter; 133} -134private String m_format; -135public boolean getFormat__IsNotDefault() +134private String m_filter; +135public boolean getFilter__IsNotDefault() 136{ -137 return m_format__IsNotDefault; +137 return m_filter__IsNotDefault; 138} -139private boolean m_format__IsNotDefault; +139private boolean m_filter__IsNotDefault; 140 } 141 @Override 142 protected org.jamon.AbstractTemplateProxy.ImplData makeImplData() @@ -156,31 +156,31 @@ 148return (ImplData) super.getImplData(); 149 } 150 -151 protected String bcv; -152 public final
[30/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html index 109b5f3..e484176 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html @@ -23,484 +23,310 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import io.netty.bootstrap.ServerBootstrap; -022import io.netty.buffer.ByteBuf; -023import io.netty.buffer.PooledByteBufAllocator; -024import io.netty.buffer.Unpooled; -025import io.netty.channel.Channel; -026import io.netty.channel.ChannelFuture; -027import io.netty.channel.ChannelFutureListener; -028import io.netty.channel.ChannelHandlerContext; -029import io.netty.channel.ChannelInboundHandlerAdapter; -030import io.netty.channel.ChannelInitializer; -031import io.netty.channel.ChannelOption; -032import io.netty.channel.ChannelOutboundHandlerAdapter; -033import io.netty.channel.ChannelPipeline; -034import io.netty.channel.ChannelPromise; -035import io.netty.channel.EventLoopGroup; -036import io.netty.channel.epoll.EpollEventLoopGroup; -037import io.netty.channel.epoll.EpollServerSocketChannel; -038import io.netty.channel.group.ChannelGroup; -039import io.netty.channel.group.DefaultChannelGroup; -040import io.netty.channel.nio.NioEventLoopGroup; -041import io.netty.channel.socket.SocketChannel; -042import io.netty.channel.socket.nio.NioServerSocketChannel; -043import io.netty.handler.codec.ByteToMessageDecoder; -044import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -045import io.netty.util.concurrent.GlobalEventExecutor; -046 -047import java.io.IOException; -048import java.io.InterruptedIOException; -049import java.net.InetAddress; -050import java.net.InetSocketAddress; -051import java.nio.ByteBuffer; -052import java.util.Arrays; -053import java.util.List; -054import java.util.concurrent.CountDownLatch; -055 -056import org.apache.commons.logging.Log; -057import org.apache.commons.logging.LogFactory; -058import org.apache.hadoop.conf.Configuration; -059import org.apache.hadoop.hbase.CellScanner; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -063import org.apache.hadoop.hbase.nio.ByteBuff; -064import org.apache.hadoop.hbase.nio.SingleByteBuff; -065import org.apache.hadoop.hbase.security.AccessDeniedException; -066import org.apache.hadoop.hbase.security.AuthMethod; -067import org.apache.hadoop.hbase.security.HBasePolicyProvider; -068import org.apache.hadoop.hbase.security.SaslStatus; -069import org.apache.hadoop.hbase.security.SaslUtil; -070import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -073import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -074import org.apache.hadoop.hbase.util.Bytes; -075import org.apache.hadoop.hbase.util.JVM; -076import org.apache.hadoop.hbase.util.Pair; -077import org.apache.hadoop.io.IntWritable; -078import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -079import org.apache.htrace.TraceInfo; +018package org.apache.hadoop.hbase.ipc; +019 +020import io.netty.bootstrap.ServerBootstrap; +021import io.netty.buffer.ByteBuf; +022import io.netty.buffer.PooledByteBufAllocator; +023import io.netty.buffer.Unpooled; +024import io.netty.channel.Channel; +025import io.netty.channel.ChannelFuture; +026import io.netty.channel.ChannelFutureListener; +027import io.netty.channel.ChannelHandlerContext; +028import io.netty.channel.ChannelInboundHandlerAdapter; +029import io.netty.channel.ChannelInitializer; +030import io.netty.channel.ChannelOption; +031import io.netty.channel.ChannelOutboundHandlerAdapter; +032import io.netty.channel.ChannelPipeline; +033import io.netty.channel.ChannelPromise; +034import io.netty.channel.EventLoopGroup; +035import io.netty.channel.epoll.EpollEventLoopGroup; +036import io.netty.channel.epoll.EpollServerSocketChannel; +037import io.netty.channel.group.ChannelGroup; +038import io.netty.channel.group.DefaultChannelGroup; +039import io.netty.channel.nio.NioEventLoopGroup; +040import io.netty.channel.socket.SocketChannel; +041import io.netty.channel.socket.nio.NioServerSocketChannel; +042import
[37/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.html b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.html new file mode 100644 index 000..eb84862 --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.html @@ -0,0 +1,716 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +SimpleServerRpcConnection (Apache HBase 2.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.ipc +Class SimpleServerRpcConnection + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.ipc.ServerRpcConnection + + +org.apache.hadoop.hbase.ipc.SimpleServerRpcConnection + + + + + + + + + +All Implemented Interfaces: +http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable, http://docs.oracle.com/javase/8/docs/api/java/lang/AutoCloseable.html?is-external=true; title="class or interface in java.lang">AutoCloseable + + + +@InterfaceAudience.Private +class SimpleServerRpcConnection +extends ServerRpcConnection +Reads calls from a connection and queues them for handling. + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +(package private) http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SocketChannel.html?is-external=true; title="class or interface in java.nio.channels">SocketChannel +channel + + +private ByteBuff +data + + +private http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBuffer +dataLengthBuffer + + +private long +lastContact + + +private SimpleRpcServerResponder +responder + + +protected http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentLinkedDeque.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentLinkedDequeSimpleServerCall +responseQueue + + +(package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">Lock +responseWriteLock + + +private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true; title="class or interface in java.util.concurrent.atomic">LongAdder +rpcCount + + +private http://docs.oracle.com/javase/8/docs/api/java/net/Socket.html?is-external=true; title="class or interface in java.net">Socket +socket + + + + + + +Fields inherited from classorg.apache.hadoop.hbase.ipc.ServerRpcConnection +addr, authenticatedWithFallback, authFailedCall, authFailedResponse, authMethod, AUTHORIZATION_FAILED_CALLID, callCleanup, codec, compressionCodec, CONNECTION_HEADER_RESPONSE_CALLID, connectionHeader, connectionHeaderRead, connectionPreambleRead, cryptoAES, hostAddress, remotePort, retryImmediatelySuppor ted, rpcServer, SASL_CALLID, saslCall, saslContextEstablished, saslServer, service, setConnectionHeaderResponseCall, skipInitialSaslHandshake, ugi, useCryptoAesWrap, user, useSasl, useWrap + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +SimpleServerRpcConnection(SimpleRpcServerrpcServer, + http://docs.oracle.com/javase/8/docs/api/java/nio/channels/SocketChannel.html?is-external=true; title="class or interface in java.nio.channels">SocketChannelchannel, + longlastContact) + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and
[06/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.html index ca92b5e..c703b08 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmpl.html @@ -69,15 +69,15 @@ 061 requiredArguments = { 062@org.jamon.annotations.Argument(name = "master", type = "HMaster")}, 063 optionalArguments = { -064@org.jamon.annotations.Argument(name = "serverManager", type = "ServerManager"), -065@org.jamon.annotations.Argument(name = "format", type = "String"), +064@org.jamon.annotations.Argument(name = "deadServers", type = "SetServerName"), +065@org.jamon.annotations.Argument(name = "frags", type = "MapString,Integer"), 066@org.jamon.annotations.Argument(name = "catalogJanitorEnabled", type = "boolean"), -067@org.jamon.annotations.Argument(name = "servers", type = "ListServerName"), -068@org.jamon.annotations.Argument(name = "filter", type = "String"), -069@org.jamon.annotations.Argument(name = "assignmentManager", type = "AssignmentManager"), -070@org.jamon.annotations.Argument(name = "deadServers", type = "SetServerName"), -071@org.jamon.annotations.Argument(name = "frags", type = "MapString,Integer"), -072@org.jamon.annotations.Argument(name = "metaLocation", type = "ServerName")}) +067@org.jamon.annotations.Argument(name = "format", type = "String"), +068@org.jamon.annotations.Argument(name = "assignmentManager", type = "AssignmentManager"), +069@org.jamon.annotations.Argument(name = "metaLocation", type = "ServerName"), +070@org.jamon.annotations.Argument(name = "filter", type = "String"), +071@org.jamon.annotations.Argument(name = "servers", type = "ListServerName"), +072@org.jamon.annotations.Argument(name = "serverManager", type = "ServerManager")}) 073public class MasterStatusTmpl 074 extends org.jamon.AbstractTemplateProxy 075{ @@ -118,40 +118,40 @@ 110 return m_master; 111} 112private HMaster m_master; -113// 28, 1 -114public void setServerManager(ServerManager serverManager) +113// 24, 1 +114public void setDeadServers(SetServerName deadServers) 115{ -116 // 28, 1 -117 m_serverManager = serverManager; -118 m_serverManager__IsNotDefault = true; +116 // 24, 1 +117 m_deadServers = deadServers; +118 m_deadServers__IsNotDefault = true; 119} -120public ServerManager getServerManager() +120public SetServerName getDeadServers() 121{ -122 return m_serverManager; +122 return m_deadServers; 123} -124private ServerManager m_serverManager; -125public boolean getServerManager__IsNotDefault() +124private SetServerName m_deadServers; +125public boolean getDeadServers__IsNotDefault() 126{ -127 return m_serverManager__IsNotDefault; +127 return m_deadServers__IsNotDefault; 128} -129private boolean m_serverManager__IsNotDefault; -130// 27, 1 -131public void setFormat(String format) +129private boolean m_deadServers__IsNotDefault; +130// 21, 1 +131public void setFrags(MapString,Integer frags) 132{ -133 // 27, 1 -134 m_format = format; -135 m_format__IsNotDefault = true; +133 // 21, 1 +134 m_frags = frags; +135 m_frags__IsNotDefault = true; 136} -137public String getFormat() +137public MapString,Integer getFrags() 138{ -139 return m_format; +139 return m_frags; 140} -141private String m_format; -142public boolean getFormat__IsNotDefault() +141private MapString,Integer m_frags; +142public boolean getFrags__IsNotDefault() 143{ -144 return m_format__IsNotDefault; +144 return m_frags__IsNotDefault; 145} -146private boolean m_format__IsNotDefault; +146private boolean m_frags__IsNotDefault; 147// 25, 1 148public void setCatalogJanitorEnabled(boolean catalogJanitorEnabled) 149{ @@ -169,108 +169,108 @@ 161 return m_catalogJanitorEnabled__IsNotDefault; 162} 163private boolean m_catalogJanitorEnabled__IsNotDefault; -164// 23, 1 -165public void setServers(ListServerName servers) +164// 27, 1 +165public void setFormat(String format) 166{ -167 // 23, 1 -168 m_servers = servers; -169 m_servers__IsNotDefault = true; +167 // 27, 1 +168 m_format = format; +169 m_format__IsNotDefault = true; 170} -171public ListServerName getServers() +171public String getFormat() 172{ -173 return m_servers; +173 return m_format; 174
[31/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html index 109b5f3..e484176 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.CallWriteListener.html @@ -23,484 +23,310 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import io.netty.bootstrap.ServerBootstrap; -022import io.netty.buffer.ByteBuf; -023import io.netty.buffer.PooledByteBufAllocator; -024import io.netty.buffer.Unpooled; -025import io.netty.channel.Channel; -026import io.netty.channel.ChannelFuture; -027import io.netty.channel.ChannelFutureListener; -028import io.netty.channel.ChannelHandlerContext; -029import io.netty.channel.ChannelInboundHandlerAdapter; -030import io.netty.channel.ChannelInitializer; -031import io.netty.channel.ChannelOption; -032import io.netty.channel.ChannelOutboundHandlerAdapter; -033import io.netty.channel.ChannelPipeline; -034import io.netty.channel.ChannelPromise; -035import io.netty.channel.EventLoopGroup; -036import io.netty.channel.epoll.EpollEventLoopGroup; -037import io.netty.channel.epoll.EpollServerSocketChannel; -038import io.netty.channel.group.ChannelGroup; -039import io.netty.channel.group.DefaultChannelGroup; -040import io.netty.channel.nio.NioEventLoopGroup; -041import io.netty.channel.socket.SocketChannel; -042import io.netty.channel.socket.nio.NioServerSocketChannel; -043import io.netty.handler.codec.ByteToMessageDecoder; -044import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -045import io.netty.util.concurrent.GlobalEventExecutor; -046 -047import java.io.IOException; -048import java.io.InterruptedIOException; -049import java.net.InetAddress; -050import java.net.InetSocketAddress; -051import java.nio.ByteBuffer; -052import java.util.Arrays; -053import java.util.List; -054import java.util.concurrent.CountDownLatch; -055 -056import org.apache.commons.logging.Log; -057import org.apache.commons.logging.LogFactory; -058import org.apache.hadoop.conf.Configuration; -059import org.apache.hadoop.hbase.CellScanner; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -063import org.apache.hadoop.hbase.nio.ByteBuff; -064import org.apache.hadoop.hbase.nio.SingleByteBuff; -065import org.apache.hadoop.hbase.security.AccessDeniedException; -066import org.apache.hadoop.hbase.security.AuthMethod; -067import org.apache.hadoop.hbase.security.HBasePolicyProvider; -068import org.apache.hadoop.hbase.security.SaslStatus; -069import org.apache.hadoop.hbase.security.SaslUtil; -070import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -073import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -074import org.apache.hadoop.hbase.util.Bytes; -075import org.apache.hadoop.hbase.util.JVM; -076import org.apache.hadoop.hbase.util.Pair; -077import org.apache.hadoop.io.IntWritable; -078import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -079import org.apache.htrace.TraceInfo; +018package org.apache.hadoop.hbase.ipc; +019 +020import io.netty.bootstrap.ServerBootstrap; +021import io.netty.buffer.ByteBuf; +022import io.netty.buffer.PooledByteBufAllocator; +023import io.netty.buffer.Unpooled; +024import io.netty.channel.Channel; +025import io.netty.channel.ChannelFuture; +026import io.netty.channel.ChannelFutureListener; +027import io.netty.channel.ChannelHandlerContext; +028import io.netty.channel.ChannelInboundHandlerAdapter; +029import io.netty.channel.ChannelInitializer; +030import io.netty.channel.ChannelOption; +031import io.netty.channel.ChannelOutboundHandlerAdapter; +032import io.netty.channel.ChannelPipeline; +033import io.netty.channel.ChannelPromise; +034import io.netty.channel.EventLoopGroup; +035import io.netty.channel.epoll.EpollEventLoopGroup; +036import io.netty.channel.epoll.EpollServerSocketChannel; +037import io.netty.channel.group.ChannelGroup; +038import io.netty.channel.group.DefaultChannelGroup; +039import io.netty.channel.nio.NioEventLoopGroup; +040import io.netty.channel.socket.SocketChannel; +041import io.netty.channel.socket.nio.NioServerSocketChannel; +042import io.netty.handler.codec.ByteToMessageDecoder; +043import
[42/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html b/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html index 434671a..ac50afc 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/ServerCall.html @@ -93,14 +93,14 @@ var activeTableTab = "activeTableTab"; org.apache.hadoop.hbase.ipc -Class ServerCall +Class ServerCallT extends ServerRpcConnection http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object -org.apache.hadoop.hbase.ipc.ServerCall +org.apache.hadoop.hbase.ipc.ServerCallT @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -abstract class ServerCall +abstract class ServerCallT extends ServerRpcConnection extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements RpcCall Datastructure that holds all necessary to a method invocation and then afterward, carries @@ -154,7 +154,7 @@ implements cellScanner -protected RpcServer.Connection +protected T connection @@ -262,13 +262,13 @@ implements Constructor and Description -ServerCall(intid, +ServerCall(intid, org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingServiceservice, org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptormd, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeaderheader, org.apache.hadoop.hbase.shaded.com.google.protobuf.Messageparam, CellScannercellScanner, - RpcServer.Connectionconnection, + Tconnection, longsize, org.apache.htrace.TraceInfotinfo, http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddressremoteAddress, @@ -528,7 +528,7 @@ implements id -protected finalint id +protected finalint id @@ -537,7 +537,7 @@ implements service -protected finalorg.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service +protected finalorg.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService service @@ -546,7 +546,7 @@ implements md -protected finalorg.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md +protected finalorg.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor md @@ -555,7 +555,7 @@ implements header -protected finalorg.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header +protected finalorg.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader header @@ -564,7 +564,7 @@ implements param -protectedorg.apache.hadoop.hbase.shaded.com.google.protobuf.Message param +protectedorg.apache.hadoop.hbase.shaded.com.google.protobuf.Message param @@ -573,7 +573,7 @@ implements cellScanner -protected finalCellScanner cellScanner +protected finalCellScanner cellScanner @@ -582,7 +582,7 @@ implements connection -protected finalRpcServer.Connection connection +protected finalT extends ServerRpcConnection connection @@ -591,7 +591,7 @@ implements receiveTime -protected finallong receiveTime +protected finallong receiveTime @@ -600,7 +600,7 @@ implements timeout -protected finalint timeout +protected finalint timeout @@ -609,7 +609,7 @@ implements startTime -protectedlong startTime +protectedlong startTime @@ -618,7 +618,7 @@ implements deadline -protected finallong deadline +protected finallong deadline @@ -627,7 +627,7 @@ implements reservoir -protected finalByteBufferPool reservoir +protected finalByteBufferPool reservoir @@ -636,7 +636,7 @@ implements cellBlockBuilder -protected finalCellBlockBuilder cellBlockBuilder +protected finalCellBlockBuilder cellBlockBuilder @@ -645,7 +645,7 @@ implements response -protectedBufferChain response +protectedBufferChain response Chain of buffers to send as response. @@ -655,7 +655,7 @@ implements size -protected finallong size +protected finallong size @@ -664,7 +664,7 @@ implements isError -protectedboolean isError +protectedboolean isError @@ -673,7 +673,7 @@ implements tinfo -protected finalorg.apache.htrace.TraceInfo tinfo +protected finalorg.apache.htrace.TraceInfo tinfo @@ -682,7 +682,7 @@ implements cellBlockStream -protectedByteBufferListOutputStream cellBlockStream +protectedByteBufferListOutputStream cellBlockStream @@ -691,7 +691,7 @@ implements reqCleanup -protectedRpcServer.CallCleanup reqCleanup +protectedRpcServer.CallCleanup
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html index 109b5f3..e484176 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html @@ -23,484 +23,310 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import io.netty.bootstrap.ServerBootstrap; -022import io.netty.buffer.ByteBuf; -023import io.netty.buffer.PooledByteBufAllocator; -024import io.netty.buffer.Unpooled; -025import io.netty.channel.Channel; -026import io.netty.channel.ChannelFuture; -027import io.netty.channel.ChannelFutureListener; -028import io.netty.channel.ChannelHandlerContext; -029import io.netty.channel.ChannelInboundHandlerAdapter; -030import io.netty.channel.ChannelInitializer; -031import io.netty.channel.ChannelOption; -032import io.netty.channel.ChannelOutboundHandlerAdapter; -033import io.netty.channel.ChannelPipeline; -034import io.netty.channel.ChannelPromise; -035import io.netty.channel.EventLoopGroup; -036import io.netty.channel.epoll.EpollEventLoopGroup; -037import io.netty.channel.epoll.EpollServerSocketChannel; -038import io.netty.channel.group.ChannelGroup; -039import io.netty.channel.group.DefaultChannelGroup; -040import io.netty.channel.nio.NioEventLoopGroup; -041import io.netty.channel.socket.SocketChannel; -042import io.netty.channel.socket.nio.NioServerSocketChannel; -043import io.netty.handler.codec.ByteToMessageDecoder; -044import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -045import io.netty.util.concurrent.GlobalEventExecutor; -046 -047import java.io.IOException; -048import java.io.InterruptedIOException; -049import java.net.InetAddress; -050import java.net.InetSocketAddress; -051import java.nio.ByteBuffer; -052import java.util.Arrays; -053import java.util.List; -054import java.util.concurrent.CountDownLatch; -055 -056import org.apache.commons.logging.Log; -057import org.apache.commons.logging.LogFactory; -058import org.apache.hadoop.conf.Configuration; -059import org.apache.hadoop.hbase.CellScanner; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -063import org.apache.hadoop.hbase.nio.ByteBuff; -064import org.apache.hadoop.hbase.nio.SingleByteBuff; -065import org.apache.hadoop.hbase.security.AccessDeniedException; -066import org.apache.hadoop.hbase.security.AuthMethod; -067import org.apache.hadoop.hbase.security.HBasePolicyProvider; -068import org.apache.hadoop.hbase.security.SaslStatus; -069import org.apache.hadoop.hbase.security.SaslUtil; -070import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -073import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -074import org.apache.hadoop.hbase.util.Bytes; -075import org.apache.hadoop.hbase.util.JVM; -076import org.apache.hadoop.hbase.util.Pair; -077import org.apache.hadoop.io.IntWritable; -078import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -079import org.apache.htrace.TraceInfo; +018package org.apache.hadoop.hbase.ipc; +019 +020import io.netty.bootstrap.ServerBootstrap; +021import io.netty.buffer.ByteBuf; +022import io.netty.buffer.PooledByteBufAllocator; +023import io.netty.buffer.Unpooled; +024import io.netty.channel.Channel; +025import io.netty.channel.ChannelFuture; +026import io.netty.channel.ChannelFutureListener; +027import io.netty.channel.ChannelHandlerContext; +028import io.netty.channel.ChannelInboundHandlerAdapter; +029import io.netty.channel.ChannelInitializer; +030import io.netty.channel.ChannelOption; +031import io.netty.channel.ChannelOutboundHandlerAdapter; +032import io.netty.channel.ChannelPipeline; +033import io.netty.channel.ChannelPromise; +034import io.netty.channel.EventLoopGroup; +035import io.netty.channel.epoll.EpollEventLoopGroup; +036import io.netty.channel.epoll.EpollServerSocketChannel; +037import io.netty.channel.group.ChannelGroup; +038import io.netty.channel.group.DefaultChannelGroup; +039import io.netty.channel.nio.NioEventLoopGroup; +040import io.netty.channel.socket.SocketChannel; +041import io.netty.channel.socket.nio.NioServerSocketChannel; +042import io.netty.handler.codec.ByteToMessageDecoder; +043import
[1/3] hbase git commit: HBASE-18012 Move RpcServer.Connection to a separated file
Repository: hbase Updated Branches: refs/heads/master 5cdaca5c0 -> 341223d86 http://git-wip-us.apache.org/repos/asf/hbase/blob/341223d8/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.java new file mode 100644 index 000..50a1a6b --- /dev/null +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/SimpleServerRpcConnection.java @@ -0,0 +1,428 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hbase.ipc; + +import java.io.IOException; +import java.io.InputStream; +import java.net.InetAddress; +import java.net.Socket; +import java.nio.ByteBuffer; +import java.nio.channels.ReadableByteChannel; +import java.nio.channels.SocketChannel; +import java.util.Arrays; +import java.util.concurrent.ConcurrentLinkedDeque; +import java.util.concurrent.atomic.LongAdder; +import java.util.concurrent.locks.Lock; +import java.util.concurrent.locks.ReentrantLock; + +import org.apache.hadoop.hbase.CellScanner; +import org.apache.hadoop.hbase.DoNotRetryIOException; +import org.apache.hadoop.hbase.HConstants; +import org.apache.hadoop.hbase.classification.InterfaceAudience; +import org.apache.hadoop.hbase.client.VersionInfoUtil; +import org.apache.hadoop.hbase.exceptions.RequestTooBigException; +import org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup; +import org.apache.hadoop.hbase.nio.ByteBuff; +import org.apache.hadoop.hbase.nio.SingleByteBuff; +import org.apache.hadoop.hbase.security.AccessDeniedException; +import org.apache.hadoop.hbase.security.AuthMethod; +import org.apache.hadoop.hbase.security.SaslStatus; +import org.apache.hadoop.hbase.security.SaslUtil; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; +import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; +import org.apache.hadoop.hbase.util.Bytes; +import org.apache.hadoop.hbase.util.Pair; +import org.apache.hadoop.io.IntWritable; +import org.apache.htrace.TraceInfo; + +/** Reads calls from a connection and queues them for handling. */ +@edu.umd.cs.findbugs.annotations.SuppressWarnings(value = "VO_VOLATILE_INCREMENT", +justification = "False positive according to http://sourceforge.net/p/findbugs/bugs/1032/;) +@InterfaceAudience.Private +class SimpleServerRpcConnection extends ServerRpcConnection { + + final SocketChannel channel; + private ByteBuff data; + private ByteBuffer dataLengthBuffer; + protected final ConcurrentLinkedDeque responseQueue = + new ConcurrentLinkedDeque<>(); + final Lock responseWriteLock = new ReentrantLock(); + private final LongAdder rpcCount = new LongAdder(); // number of outstanding rpcs + private long lastContact; + private final Socket socket; + private final SimpleRpcServerResponder responder; + + public SimpleServerRpcConnection(SimpleRpcServer rpcServer, SocketChannel channel, + long lastContact) { +super(rpcServer); +this.channel = channel; +this.lastContact = lastContact; +this.data = null; +this.dataLengthBuffer = ByteBuffer.allocate(4); +this.socket = channel.socket(); +this.addr = socket.getInetAddress(); +if (addr == null) { + this.hostAddress = "*Unknown*"; +} else { + this.hostAddress = addr.getHostAddress(); +} +this.remotePort = socket.getPort(); +if (rpcServer.socketSendBufferSize != 0) { + try { +socket.setSendBufferSize(rpcServer.socketSendBufferSize); + } catch (IOException e) { +SimpleRpcServer.LOG.warn( + "Connection: unable to set socket send buffer size to " + rpcServer.socketSendBufferSize); + } +} +this.saslCall = new
[2/3] hbase git commit: HBASE-18012 Move RpcServer.Connection to a separated file
http://git-wip-us.apache.org/repos/asf/hbase/blob/341223d8/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcConnection.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcConnection.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcConnection.java new file mode 100644 index 000..d4ab95c --- /dev/null +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcConnection.java @@ -0,0 +1,852 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hbase.ipc; + +import java.io.ByteArrayInputStream; +import java.io.ByteArrayOutputStream; +import java.io.Closeable; +import java.io.DataOutputStream; +import java.io.IOException; +import java.net.InetAddress; +import java.net.InetSocketAddress; +import java.nio.ByteBuffer; +import java.nio.channels.Channels; +import java.nio.channels.ReadableByteChannel; +import java.security.GeneralSecurityException; +import java.security.PrivilegedExceptionAction; +import java.util.Properties; + +import javax.security.sasl.Sasl; +import javax.security.sasl.SaslException; +import javax.security.sasl.SaslServer; + +import org.apache.commons.crypto.cipher.CryptoCipherFactory; +import org.apache.commons.crypto.random.CryptoRandom; +import org.apache.commons.crypto.random.CryptoRandomFactory; +import org.apache.hadoop.hbase.CellScanner; +import org.apache.hadoop.hbase.DoNotRetryIOException; +import org.apache.hadoop.hbase.client.VersionInfoUtil; +import org.apache.hadoop.hbase.codec.Codec; +import org.apache.hadoop.hbase.io.ByteBufferOutputStream; +import org.apache.hadoop.hbase.io.crypto.aes.CryptoAES; +import org.apache.hadoop.hbase.ipc.RpcServer.CallCleanup; +import org.apache.hadoop.hbase.nio.ByteBuff; +import org.apache.hadoop.hbase.nio.SingleByteBuff; +import org.apache.hadoop.hbase.security.AccessDeniedException; +import org.apache.hadoop.hbase.security.AuthMethod; +import org.apache.hadoop.hbase.security.HBaseSaslRpcServer; +import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslDigestCallbackHandler; +import org.apache.hadoop.hbase.security.HBaseSaslRpcServer.SaslGssCallbackHandler; +import org.apache.hadoop.hbase.security.SaslStatus; +import org.apache.hadoop.hbase.security.SaslUtil; +import org.apache.hadoop.hbase.security.User; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteInput; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; +import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; +import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; +import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; +import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos; +import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader; +import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; +import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation; +import org.apache.hadoop.hbase.util.Bytes; +import org.apache.hadoop.io.BytesWritable; +import org.apache.hadoop.io.Writable; +import org.apache.hadoop.io.WritableUtils; +import org.apache.hadoop.io.compress.CompressionCodec; +import org.apache.hadoop.security.UserGroupInformation; +import org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod; +import org.apache.hadoop.security.authorize.AuthorizationException; +import org.apache.hadoop.security.authorize.ProxyUsers; +import org.apache.hadoop.security.token.SecretManager.InvalidToken; +import org.apache.hadoop.security.token.TokenIdentifier; +import org.apache.htrace.TraceInfo; + +/** Reads calls from a connection and queues them for handling. */
[3/3] hbase git commit: HBASE-18012 Move RpcServer.Connection to a separated file
HBASE-18012 Move RpcServer.Connection to a separated file Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/341223d8 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/341223d8 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/341223d8 Branch: refs/heads/master Commit: 341223d86c567e2124802b0a19e1ba7bdc560cad Parents: 5cdaca5 Author: zhangduoAuthored: Wed May 10 11:05:38 2017 +0800 Committer: zhangduo Committed: Mon May 15 18:07:38 2017 +0800 -- .../apache/hadoop/hbase/nio/SingleByteBuff.java | 4 +- .../org/apache/hadoop/hbase/ipc/CallRunner.java | 4 +- .../apache/hadoop/hbase/ipc/NettyRpcServer.java | 186 +--- .../hadoop/hbase/ipc/NettyServerCall.java | 13 +- .../hbase/ipc/NettyServerRpcConnection.java | 206 + .../org/apache/hadoop/hbase/ipc/RpcServer.java | 820 +- .../org/apache/hadoop/hbase/ipc/ServerCall.java | 7 +- .../hadoop/hbase/ipc/ServerRpcConnection.java | 852 +++ .../hadoop/hbase/ipc/SimpleRpcServer.java | 717 +--- .../hbase/ipc/SimpleRpcServerResponder.java | 316 +++ .../hadoop/hbase/ipc/SimpleServerCall.java | 18 +- .../hbase/ipc/SimpleServerRpcConnection.java| 428 ++ .../hbase/security/HBaseSaslRpcServer.java | 24 +- .../hadoop/hbase/ipc/AbstractTestIPC.java | 11 +- 14 files changed, 1869 insertions(+), 1737 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/341223d8/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java index 9f6b7b5..6ed90ad 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java @@ -17,6 +17,8 @@ */ package org.apache.hadoop.hbase.nio; +import com.google.common.annotations.VisibleForTesting; + import java.io.IOException; import java.nio.ByteBuffer; import java.nio.channels.ReadableByteChannel; @@ -27,8 +29,6 @@ import org.apache.hadoop.hbase.util.ObjectIntPair; import org.apache.hadoop.hbase.util.UnsafeAccess; import org.apache.hadoop.hbase.util.UnsafeAvailChecker; -import com.google.common.annotations.VisibleForTesting; - import sun.nio.ch.DirectBuffer; /** http://git-wip-us.apache.org/repos/asf/hbase/blob/341223d8/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java index f16fc50..f476b11 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/CallRunner.java @@ -75,8 +75,8 @@ public class CallRunner { * @deprecated As of release 2.0, this will be removed in HBase 3.0 */ @Deprecated - public ServerCall getCall() { -return (ServerCall) call; + public ServerCall getCall() { +return (ServerCall) call; } public void setStatus(MonitoredRPCHandler status) { http://git-wip-us.apache.org/repos/asf/hbase/blob/341223d8/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java index c18b894..4a4ddba 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcServer.java @@ -15,7 +15,6 @@ * See the License for the specific language governing permissions and * limitations under the License. */ - package org.apache.hadoop.hbase.ipc; import io.netty.bootstrap.ServerBootstrap; @@ -46,10 +45,7 @@ import io.netty.util.concurrent.GlobalEventExecutor; import java.io.IOException; import java.io.InterruptedIOException; -import java.net.InetAddress; import java.net.InetSocketAddress; -import java.nio.ByteBuffer; -import java.util.Arrays; import java.util.List; import java.util.concurrent.CountDownLatch; @@ -57,31 +53,21 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.CellScanner; -import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.Server;