[11/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServer.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServer.html index ab92b9d..99959bb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftServer.html @@ -149,655 +149,666 @@ 141import org.slf4j.Logger; 142import org.slf4j.LoggerFactory; 143 -144import org.apache.hbase.thirdparty.com.google.common.base.Joiner; -145import org.apache.hbase.thirdparty.com.google.common.base.Splitter; -146import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -147import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine; -148import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLineParser; -149import org.apache.hbase.thirdparty.org.apache.commons.cli.DefaultParser; -150import org.apache.hbase.thirdparty.org.apache.commons.cli.HelpFormatter; -151import org.apache.hbase.thirdparty.org.apache.commons.cli.Options; -152 -153/** -154 * ThriftServer- this class starts up a Thrift server which implements the -155 * Hbase API specified in the Hbase.thrift IDL file. The server runs in an -156 * independent process. -157 */ -158@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -159public class ThriftServer extends Configured implements Tool { -160 -161 private static final Logger LOG = LoggerFactory.getLogger(ThriftServer.class); -162 +144import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +145import org.apache.hbase.thirdparty.com.google.common.base.Joiner; +146import org.apache.hbase.thirdparty.com.google.common.base.Splitter; +147import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; +148import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine; +149import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLineParser; +150import org.apache.hbase.thirdparty.org.apache.commons.cli.DefaultParser; +151import org.apache.hbase.thirdparty.org.apache.commons.cli.HelpFormatter; +152import org.apache.hbase.thirdparty.org.apache.commons.cli.Options; +153 +154/** +155 * ThriftServer- this class starts up a Thrift server which implements the +156 * Hbase API specified in the Hbase.thrift IDL file. The server runs in an +157 * independent process. +158 */ +159@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) +160public class ThriftServer extends Configured implements Tool { +161 +162 private static final Logger LOG = LoggerFactory.getLogger(ThriftServer.class); 163 164 -165 protected Configuration conf; -166 -167 protected InfoServer infoServer; -168 -169 protected TProcessor processor; -170 -171 protected ThriftMetrics metrics; -172 protected HBaseServiceHandler hbaseServiceHandler; -173 protected UserGroupInformation serviceUGI; -174 protected boolean httpEnabled; -175 -176 protected SaslUtil.QualityOfProtection qop; -177 protected String host; -178 protected int listenPort; -179 +165 +166 protected Configuration conf; +167 +168 protected InfoServer infoServer; +169 +170 protected TProcessor processor; +171 +172 protected ThriftMetrics metrics; +173 protected HBaseServiceHandler hbaseServiceHandler; +174 protected UserGroupInformation serviceUGI; +175 protected boolean httpEnabled; +176 +177 protected SaslUtil.QualityOfProtection qop; +178 protected String host; +179 protected int listenPort; 180 -181 protected boolean securityEnabled; -182 protected boolean doAsEnabled; -183 -184 protected JvmPauseMonitor pauseMonitor; -185 -186 protected volatile TServer tserver; -187 protected volatile Server httpServer; -188 +181 +182 protected boolean securityEnabled; +183 protected boolean doAsEnabled; +184 +185 protected JvmPauseMonitor pauseMonitor; +186 +187 protected volatile TServer tserver; +188 protected volatile Server httpServer; 189 -190 // -191 // Main program and support routines -192 // -193 -194 public ThriftServer(Configuration conf) { -195this.conf = HBaseConfiguration.create(conf); -196 } -197 -198 protected void setupParamters() throws IOException { -199// login the server principal (if using secure Hadoop) -200UserProvider userProvider = UserProvider.instantiate(conf); -201securityEnabled = userProvider.isHadoopSecurityEnabled() -202 userProvider.isHBaseSecurityEnabled(); -203if (securityEnabled) { -204 host = Strings.domainNamePointerToHostName(DNS.getDefaultHost( -205 conf.get(THRIFT_DNS_INTERFACE_KEY, "default"), -206 conf.get(THRIFT_DNS_NAMESERVER_KEY, "default"))); -207 userProvider.login(THRIFT_KEYTAB_FILE_KEY,
[11/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureStateSerializer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureStateSerializer.html b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureStateSerializer.html index 4eaf179..6fc7bf8 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureStateSerializer.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureStateSerializer.html @@ -266,12 +266,20 @@ protected void -CreateTableProcedure.deserializeStateData(ProcedureStateSerializerserializer) +SwitchRpcThrottleRemoteProcedure.deserializeStateData(ProcedureStateSerializerserializer) protected void +CreateTableProcedure.deserializeStateData(ProcedureStateSerializerserializer) + + +protected void ModifyNamespaceProcedure.deserializeStateData(ProcedureStateSerializerserializer) + +protected void +SplitWALRemoteProcedure.deserializeStateData(ProcedureStateSerializerserializer) + protected void TruncateTableProcedure.deserializeStateData(ProcedureStateSerializerserializer) @@ -306,16 +314,24 @@ protected void -ModifyTableProcedure.deserializeStateData(ProcedureStateSerializerserializer) +SplitWALProcedure.deserializeStateData(ProcedureStateSerializerserializer) protected void -InitMetaProcedure.deserializeStateData(ProcedureStateSerializerserializer) +ModifyTableProcedure.deserializeStateData(ProcedureStateSerializerserializer) protected void +InitMetaProcedure.deserializeStateData(ProcedureStateSerializerserializer) + + +protected void ReopenTableRegionsProcedure.deserializeStateData(ProcedureStateSerializerserializer) + +protected void +SwitchRpcThrottleProcedure.deserializeStateData(ProcedureStateSerializerserializer) + protected void DeleteNamespaceProcedure.deserializeStateData(ProcedureStateSerializerserializer) @@ -332,12 +348,20 @@ protected void -CreateTableProcedure.serializeStateData(ProcedureStateSerializerserializer) +SwitchRpcThrottleRemoteProcedure.serializeStateData(ProcedureStateSerializerserializer) protected void +CreateTableProcedure.serializeStateData(ProcedureStateSerializerserializer) + + +protected void ModifyNamespaceProcedure.serializeStateData(ProcedureStateSerializerserializer) + +protected void +SplitWALRemoteProcedure.serializeStateData(ProcedureStateSerializerserializer) + protected void TruncateTableProcedure.serializeStateData(ProcedureStateSerializerserializer) @@ -372,16 +396,24 @@ protected void -ModifyTableProcedure.serializeStateData(ProcedureStateSerializerserializer) +SplitWALProcedure.serializeStateData(ProcedureStateSerializerserializer) protected void -InitMetaProcedure.serializeStateData(ProcedureStateSerializerserializer) +ModifyTableProcedure.serializeStateData(ProcedureStateSerializerserializer) protected void +InitMetaProcedure.serializeStateData(ProcedureStateSerializerserializer) + + +protected void ReopenTableRegionsProcedure.serializeStateData(ProcedureStateSerializerserializer) + +protected void +SwitchRpcThrottleProcedure.serializeStateData(ProcedureStateSerializerserializer) + protected void DeleteNamespaceProcedure.serializeStateData(ProcedureStateSerializerserializer) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html index 0aeeb85..7f8754f 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html @@ -83,22 +83,26 @@ -org.apache.hadoop.hbase.master.assignment +org.apache.hadoop.hbase.master -org.apache.hadoop.hbase.master.locking +org.apache.hadoop.hbase.master.assignment -org.apache.hadoop.hbase.master.procedure +org.apache.hadoop.hbase.master.locking -org.apache.hadoop.hbase.master.replication +org.apache.hadoop.hbase.master.procedure +org.apache.hadoop.hbase.master.replication + + + org.apache.hadoop.hbase.procedure2 @@ -107,6 +111,26 @@ + + + +Uses of ProcedureSuspendedException in org.apache.hadoop.hbase.master + +Methods in org.apache.hadoop.hbase.master that throw ProcedureSuspendedException + +Modifier and Type +Method and Description + + + +ServerName +SplitWALManager.acquireSplitWALWorker(Procedure?procedure) +try to acquire an worker from online servers which is executring + + + + + @@ -200,6 +224,19 @@ Uses of ProcedureSuspendedException in
[11/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html index 736388b..197b99d 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html @@ -26,3624 +26,3599 @@ 018package org.apache.hadoop.hbase.client; 019 020import static org.apache.hadoop.hbase.TableName.META_TABLE_NAME; -021 -022import com.google.protobuf.Message; -023import com.google.protobuf.RpcChannel; -024import java.io.IOException; -025import java.util.ArrayList; -026import java.util.Arrays; -027import java.util.Collections; -028import java.util.EnumSet; -029import java.util.HashMap; -030import java.util.List; -031import java.util.Map; -032import java.util.Optional; -033import java.util.Set; -034import java.util.concurrent.CompletableFuture; -035import java.util.concurrent.ConcurrentHashMap; -036import java.util.concurrent.TimeUnit; -037import java.util.concurrent.atomic.AtomicReference; -038import java.util.function.BiConsumer; -039import java.util.function.Function; -040import java.util.function.Supplier; -041import java.util.regex.Pattern; -042import java.util.stream.Collectors; -043import java.util.stream.Stream; -044import org.apache.commons.io.IOUtils; -045import org.apache.hadoop.conf.Configuration; -046import org.apache.hadoop.hbase.AsyncMetaTableAccessor; -047import org.apache.hadoop.hbase.CacheEvictionStats; -048import org.apache.hadoop.hbase.CacheEvictionStatsAggregator; -049import org.apache.hadoop.hbase.ClusterMetrics; -050import org.apache.hadoop.hbase.ClusterMetrics.Option; -051import org.apache.hadoop.hbase.ClusterMetricsBuilder; -052import org.apache.hadoop.hbase.HConstants; -053import org.apache.hadoop.hbase.HRegionLocation; -054import org.apache.hadoop.hbase.MetaTableAccessor; -055import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -056import org.apache.hadoop.hbase.NamespaceDescriptor; -057import org.apache.hadoop.hbase.RegionLocations; -058import org.apache.hadoop.hbase.RegionMetrics; -059import org.apache.hadoop.hbase.RegionMetricsBuilder; -060import org.apache.hadoop.hbase.ServerName; -061import org.apache.hadoop.hbase.TableExistsException; -062import org.apache.hadoop.hbase.TableName; -063import org.apache.hadoop.hbase.TableNotDisabledException; -064import org.apache.hadoop.hbase.TableNotEnabledException; -065import org.apache.hadoop.hbase.TableNotFoundException; -066import org.apache.hadoop.hbase.UnknownRegionException; -067import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder; -068import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder; -069import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.ServerRequestCallerBuilder; -070import org.apache.hadoop.hbase.client.Scan.ReadType; -071import org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil; -072import org.apache.hadoop.hbase.client.replication.TableCFs; -073import org.apache.hadoop.hbase.client.security.SecurityCapability; -074import org.apache.hadoop.hbase.exceptions.DeserializationException; -075import org.apache.hadoop.hbase.ipc.HBaseRpcController; -076import org.apache.hadoop.hbase.quotas.QuotaFilter; -077import org.apache.hadoop.hbase.quotas.QuotaSettings; -078import org.apache.hadoop.hbase.quotas.QuotaTableUtil; -079import org.apache.hadoop.hbase.replication.ReplicationException; -080import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -081import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -082import org.apache.hadoop.hbase.replication.SyncReplicationState; -083import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; -084import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException; -085import org.apache.hadoop.hbase.snapshot.SnapshotCreationException; -086import org.apache.hadoop.hbase.util.Bytes; -087import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -088import org.apache.hadoop.hbase.util.ForeignExceptionUtil; -089import org.apache.yetus.audience.InterfaceAudience; -090import org.slf4j.Logger; -091import org.slf4j.LoggerFactory; -092 -093import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -094import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; -095import org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback; -096import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; -097import org.apache.hbase.thirdparty.io.netty.util.Timeout; -098import
[11/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html deleted file mode 100644 index 1a0f64e..000 --- a/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html +++ /dev/null @@ -1,935 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - -Source code - - - - -001/** -002 * -003 * Licensed to the Apache Software Foundation (ASF) under one -004 * or more contributor license agreements. See the NOTICE file -005 * distributed with this work for additional information -006 * regarding copyright ownership. The ASF licenses this file -007 * to you under the Apache License, Version 2.0 (the -008 * "License"); you may not use this file except in compliance -009 * with the License. You may obtain a copy of the License at -010 * -011 * http://www.apache.org/licenses/LICENSE-2.0 -012 * -013 * Unless required by applicable law or agreed to in writing, software -014 * distributed under the License is distributed on an "AS IS" BASIS, -015 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -016 * See the License for the specific language governing permissions and -017 * limitations under the License. -018 */ -019package org.apache.hadoop.hbase.thrift2; -020 -021import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.appendFromThrift; -022import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.columnFamilyDescriptorFromThrift; -023import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.compareOpFromThrift; -024import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.deleteFromThrift; -025import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.deletesFromThrift; -026import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.getFromThrift; -027import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.getsFromThrift; -028import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.incrementFromThrift; -029import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.namespaceDescriptorFromHBase; -030import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.namespaceDescriptorFromThrift; -031import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.namespaceDescriptorsFromHBase; -032import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.putFromThrift; -033import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.putsFromThrift; -034import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.resultFromHBase; -035import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.resultsFromHBase; -036import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.rowMutationsFromThrift; -037import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.scanFromThrift; -038import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.splitKeyFromThrift; -039import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.tableDescriptorFromHBase; -040import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.tableDescriptorFromThrift; -041import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.tableDescriptorsFromHBase; -042import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.tableNameFromThrift; -043import static org.apache.hadoop.hbase.thrift2.ThriftUtilities.tableNamesFromHBase; -044import static org.apache.thrift.TBaseHelper.byteBufferToByteArray; -045 -046import java.io.IOException; -047import java.lang.reflect.InvocationHandler; -048import java.lang.reflect.InvocationTargetException; -049import java.lang.reflect.Method; -050import java.lang.reflect.Proxy; -051import java.nio.ByteBuffer; -052import java.util.ArrayList; -053import java.util.Collections; -054import java.util.List; -055import java.util.Map; -056import java.util.concurrent.ConcurrentHashMap; -057import java.util.concurrent.atomic.AtomicInteger; -058import java.util.regex.Pattern; -059 -060import org.apache.hadoop.conf.Configuration; -061import org.apache.hadoop.hbase.DoNotRetryIOException; -062import org.apache.hadoop.hbase.HRegionLocation; -063import org.apache.hadoop.hbase.NamespaceDescriptor; -064import org.apache.hadoop.hbase.TableName; -065import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -066import org.apache.hadoop.hbase.client.RegionLocator; -067import org.apache.hadoop.hbase.client.ResultScanner; -068import org.apache.hadoop.hbase.client.Table; -069import org.apache.hadoop.hbase.client.TableDescriptor; -070import org.apache.hadoop.hbase.security.UserProvider; -071import org.apache.hadoop.hbase.thrift.ThriftMetrics;
[11/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.OutputStreamSupplier.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.OutputStreamSupplier.html b/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.OutputStreamSupplier.html index e39bb4d..d353d02 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.OutputStreamSupplier.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.OutputStreamSupplier.html @@ -238,6 +238,6 @@ var activeTableTab = "activeTableTab"; -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.html b/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.html index 00bf3c1..d3780f5 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/CellBlockBuilder.html @@ -626,6 +626,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/ipc/CellScannerButNoCodecException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/CellScannerButNoCodecException.html b/devapidocs/org/apache/hadoop/hbase/ipc/CellScannerButNoCodecException.html index e48a2b8..df3c69d 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/CellScannerButNoCodecException.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/CellScannerButNoCodecException.html @@ -270,6 +270,6 @@ extends -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/ipc/ConnectionId.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/ConnectionId.html b/devapidocs/org/apache/hadoop/hbase/ipc/ConnectionId.html index 431190b..47aba39 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/ConnectionId.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/ConnectionId.html @@ -447,6 +447,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcChannel.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcChannel.html b/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcChannel.html index 6cc89d7..062c301 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcChannel.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcChannel.html @@ -205,6 +205,6 @@ extends com.google.protobuf.RpcChannel, com.google.protobuf.BlockingRpcChannel -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcUtils.BlockingRpcCallback.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcUtils.BlockingRpcCallback.html b/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcUtils.BlockingRpcCallback.html index 63d7ff1..d0913d0 100644 --- a/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcUtils.BlockingRpcCallback.html +++ b/devapidocs/org/apache/hadoop/hbase/ipc/CoprocessorRpcUtils.BlockingRpcCallback.html @@ -369,6 +369,6 @@ implements com.google.protobuf.RpcCallbackR -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved.
[11/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html index 0f5a095..50bf692 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html @@ -78,8712 +78,8714 @@ 070import java.util.concurrent.locks.ReadWriteLock; 071import java.util.concurrent.locks.ReentrantReadWriteLock; 072import java.util.function.Function; -073import org.apache.hadoop.conf.Configuration; -074import org.apache.hadoop.fs.FileStatus; -075import org.apache.hadoop.fs.FileSystem; -076import org.apache.hadoop.fs.LocatedFileStatus; -077import org.apache.hadoop.fs.Path; -078import org.apache.hadoop.hbase.Cell; -079import org.apache.hadoop.hbase.CellBuilderType; -080import org.apache.hadoop.hbase.CellComparator; -081import org.apache.hadoop.hbase.CellComparatorImpl; -082import org.apache.hadoop.hbase.CellScanner; -083import org.apache.hadoop.hbase.CellUtil; -084import org.apache.hadoop.hbase.CompareOperator; -085import org.apache.hadoop.hbase.CompoundConfiguration; -086import org.apache.hadoop.hbase.DoNotRetryIOException; -087import org.apache.hadoop.hbase.DroppedSnapshotException; -088import org.apache.hadoop.hbase.ExtendedCellBuilderFactory; -089import org.apache.hadoop.hbase.HConstants; -090import org.apache.hadoop.hbase.HConstants.OperationStatusCode; -091import org.apache.hadoop.hbase.HDFSBlocksDistribution; -092import org.apache.hadoop.hbase.KeyValue; -093import org.apache.hadoop.hbase.KeyValueUtil; -094import org.apache.hadoop.hbase.NamespaceDescriptor; -095import org.apache.hadoop.hbase.NotServingRegionException; -096import org.apache.hadoop.hbase.PrivateCellUtil; -097import org.apache.hadoop.hbase.RegionTooBusyException; -098import org.apache.hadoop.hbase.Tag; -099import org.apache.hadoop.hbase.TagUtil; -100import org.apache.hadoop.hbase.UnknownScannerException; -101import org.apache.hadoop.hbase.client.Append; -102import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -103import org.apache.hadoop.hbase.client.CompactionState; -104import org.apache.hadoop.hbase.client.Delete; -105import org.apache.hadoop.hbase.client.Durability; -106import org.apache.hadoop.hbase.client.Get; -107import org.apache.hadoop.hbase.client.Increment; -108import org.apache.hadoop.hbase.client.IsolationLevel; -109import org.apache.hadoop.hbase.client.Mutation; -110import org.apache.hadoop.hbase.client.PackagePrivateFieldAccessor; -111import org.apache.hadoop.hbase.client.Put; -112import org.apache.hadoop.hbase.client.RegionInfo; -113import org.apache.hadoop.hbase.client.RegionInfoBuilder; -114import org.apache.hadoop.hbase.client.RegionReplicaUtil; -115import org.apache.hadoop.hbase.client.Result; -116import org.apache.hadoop.hbase.client.RowMutations; -117import org.apache.hadoop.hbase.client.Scan; -118import org.apache.hadoop.hbase.client.TableDescriptor; -119import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -120import org.apache.hadoop.hbase.conf.ConfigurationManager; -121import org.apache.hadoop.hbase.conf.PropagatingConfigurationObserver; -122import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -123import org.apache.hadoop.hbase.coprocessor.RegionObserver.MutationType; -124import org.apache.hadoop.hbase.errorhandling.ForeignExceptionSnare; -125import org.apache.hadoop.hbase.exceptions.FailedSanityCheckException; -126import org.apache.hadoop.hbase.exceptions.TimeoutIOException; -127import org.apache.hadoop.hbase.exceptions.UnknownProtocolException; -128import org.apache.hadoop.hbase.filter.ByteArrayComparable; -129import org.apache.hadoop.hbase.filter.FilterWrapper; -130import org.apache.hadoop.hbase.filter.IncompatibleFilterException; -131import org.apache.hadoop.hbase.io.HFileLink; -132import org.apache.hadoop.hbase.io.HeapSize; -133import org.apache.hadoop.hbase.io.TimeRange; -134import org.apache.hadoop.hbase.io.hfile.BlockCache; -135import org.apache.hadoop.hbase.io.hfile.HFile; -136import org.apache.hadoop.hbase.ipc.CallerDisconnectedException; -137import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -138import org.apache.hadoop.hbase.ipc.RpcCall; -139import org.apache.hadoop.hbase.ipc.RpcServer; -140import org.apache.hadoop.hbase.mob.MobFileCache; -141import org.apache.hadoop.hbase.monitoring.MonitoredTask; -142import org.apache.hadoop.hbase.monitoring.TaskMonitor; -143import org.apache.hadoop.hbase.quotas.RegionServerSpaceQuotaManager; -144import org.apache.hadoop.hbase.regionserver.MultiVersionConcurrencyControl.WriteEntry; -145import
[11/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.ByteArrayKeyBlockIndexReader.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.ByteArrayKeyBlockIndexReader.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.ByteArrayKeyBlockIndexReader.html index 1124f8b..0c29054 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.ByteArrayKeyBlockIndexReader.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.ByteArrayKeyBlockIndexReader.html @@ -,587 +,592 @@ 1103 blockStream.write(midKeyMetadata); 1104 blockWriter.writeHeaderAndData(out); 1105if (cacheConf != null) { -1106 HFileBlock blockForCaching = blockWriter.getBlockForCaching(cacheConf); -1107 cacheConf.getBlockCache().cacheBlock(new BlockCacheKey(nameForCaching, -1108rootLevelIndexPos, true, blockForCaching.getBlockType()), blockForCaching); -1109} -1110 } - -1112 // Add root index block size -1113 totalBlockOnDiskSize += blockWriter.getOnDiskSizeWithoutHeader(); -1114 totalBlockUncompressedSize += -1115 blockWriter.getUncompressedSizeWithoutHeader(); -1116 -1117 if (LOG.isTraceEnabled()) { -1118LOG.trace("Wrote a " + numLevels + "-level index with root level at pos " -1119 + rootLevelIndexPos + ", " + rootChunk.getNumEntries() -1120 + " root-level entries, " + totalNumEntries + " total entries, " -1121 + StringUtils.humanReadableInt(this.totalBlockOnDiskSize) + -1122 " on-disk size, " -1123 + StringUtils.humanReadableInt(totalBlockUncompressedSize) + -1124 " total uncompressed size."); -1125 } -1126 return rootLevelIndexPos; -1127} -1128 -1129/** -1130 * Writes the block index data as a single level only. Does not do any -1131 * block framing. -1132 * -1133 * @param out the buffered output stream to write the index to. Typically a -1134 * stream writing into an {@link HFile} block. -1135 * @param description a short description of the index being written. Used -1136 * in a log message. -1137 * @throws IOException -1138 */ -1139public void writeSingleLevelIndex(DataOutput out, String description) -1140throws IOException { -1141 expectNumLevels(1); -1142 -1143 if (!singleLevelOnly) -1144throw new IOException("Single-level mode is turned off"); -1145 -1146 if (rootChunk.getNumEntries() 0) -1147throw new IOException("Root-level entries already added in " + -1148"single-level mode"); -1149 -1150 rootChunk = curInlineChunk; -1151 curInlineChunk = new BlockIndexChunk(); -1152 -1153 if (LOG.isTraceEnabled()) { -1154LOG.trace("Wrote a single-level " + description + " index with " -1155 + rootChunk.getNumEntries() + " entries, " + rootChunk.getRootSize() -1156 + " bytes"); -1157 } -1158 rootChunk.writeRoot(out); -1159} -1160 -1161/** -1162 * Split the current level of the block index into intermediate index -1163 * blocks of permitted size and write those blocks to disk. Return the next -1164 * level of the block index referencing those intermediate-level blocks. -1165 * -1166 * @param out -1167 * @param currentLevel the current level of the block index, such as the a -1168 * chunk referencing all leaf-level index blocks -1169 * @return the parent level block index, which becomes the root index after -1170 * a few (usually zero) iterations -1171 * @throws IOException -1172 */ -1173private BlockIndexChunk writeIntermediateLevel(FSDataOutputStream out, -1174BlockIndexChunk currentLevel) throws IOException { -1175 // Entries referencing intermediate-level blocks we are about to create. -1176 BlockIndexChunk parent = new BlockIndexChunk(); -1177 -1178 // The current intermediate-level block index chunk. -1179 BlockIndexChunk curChunk = new BlockIndexChunk(); -1180 -1181 for (int i = 0; i currentLevel.getNumEntries(); ++i) { -1182 curChunk.add(currentLevel.getBlockKey(i), -1183 currentLevel.getBlockOffset(i), currentLevel.getOnDiskDataSize(i)); -1184 -1185// HBASE-16288: We have to have at least minIndexNumEntries(16) items in the index so that -1186// we won't end up with too-many levels for a index with very large rowKeys. Also, if the -1187// first key is larger than maxChunkSize this will cause infinite recursion. -1188if (i = minIndexNumEntries curChunk.getRootSize() = maxChunkSize) { -1189 writeIntermediateBlock(out, parent, curChunk); -1190
[11/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html index a957d31..62f81b6 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html @@ -142,5192 +142,5186 @@ 134import org.apache.hadoop.hbase.wal.WAL; 135import org.apache.hadoop.hbase.wal.WALFactory; 136import org.apache.hadoop.hbase.wal.WALSplitter; -137import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; -138import org.apache.hadoop.hbase.zookeeper.ZKUtil; -139import org.apache.hadoop.hbase.zookeeper.ZKWatcher; -140import org.apache.hadoop.hbase.zookeeper.ZNodePaths; -141import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException; -142import org.apache.hadoop.ipc.RemoteException; -143import org.apache.hadoop.security.UserGroupInformation; -144import org.apache.hadoop.util.ReflectionUtils; -145import org.apache.hadoop.util.Tool; -146import org.apache.hadoop.util.ToolRunner; -147import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -148import org.apache.hbase.thirdparty.com.google.common.collect.Sets; -149import org.apache.yetus.audience.InterfaceAudience; -150import org.apache.yetus.audience.InterfaceStability; -151import org.apache.zookeeper.KeeperException; -152import org.slf4j.Logger; -153import org.slf4j.LoggerFactory; -154 -155import org.apache.hbase.thirdparty.com.google.common.base.Joiner; -156import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; -157import org.apache.hbase.thirdparty.com.google.common.collect.ImmutableList; -158import org.apache.hbase.thirdparty.com.google.common.collect.Lists; -159import org.apache.hbase.thirdparty.com.google.common.collect.Multimap; -160import org.apache.hbase.thirdparty.com.google.common.collect.Ordering; -161import org.apache.hbase.thirdparty.com.google.common.collect.TreeMultimap; -162 -163import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -164import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface; -165 -166/** -167 * HBaseFsck (hbck) is a tool for checking and repairing region consistency and -168 * table integrity problems in a corrupted HBase. This tool was written for hbase-1.x. It does not -169 * work with hbase-2.x; it can read state but is not allowed to change state; i.e. effect 'repair'. -170 * See hbck2 (HBASE-19121) for a hbck tool for hbase2. -171 * -172 * p -173 * Region consistency checks verify that hbase:meta, region deployment on region -174 * servers and the state of data in HDFS (.regioninfo files) all are in -175 * accordance. -176 * p -177 * Table integrity checks verify that all possible row keys resolve to exactly -178 * one region of a table. This means there are no individual degenerate -179 * or backwards regions; no holes between regions; and that there are no -180 * overlapping regions. -181 * p -182 * The general repair strategy works in two phases: -183 * ol -184 * li Repair Table Integrity on HDFS. (merge or fabricate regions) -185 * li Repair Region Consistency with hbase:meta and assignments -186 * /ol -187 * p -188 * For table integrity repairs, the tables' region directories are scanned -189 * for .regioninfo files. Each table's integrity is then verified. If there -190 * are any orphan regions (regions with no .regioninfo files) or holes, new -191 * regions are fabricated. Backwards regions are sidelined as well as empty -192 * degenerate (endkey==startkey) regions. If there are any overlapping regions, -193 * a new region is created and all data is merged into the new region. -194 * p -195 * Table integrity repairs deal solely with HDFS and could potentially be done -196 * offline -- the hbase region servers or master do not need to be running. -197 * This phase can eventually be used to completely reconstruct the hbase:meta table in -198 * an offline fashion. -199 * p -200 * Region consistency requires three conditions -- 1) valid .regioninfo file -201 * present in an HDFS region dir, 2) valid row with .regioninfo data in META, -202 * and 3) a region is deployed only at the regionserver that was assigned to -203 * with proper state in the master. -204 * p -205 * Region consistency repairs require hbase to be online so that hbck can -206 * contact the HBase master and region servers. The hbck#connect() method must -207 * first be called successfully. Much of the region consistency information -208 * is transient and less risky to repair. -209 * p -210 * If hbck is run from the command line, there are a handful of arguments that -211 * can be used to limit
[11/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html b/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html index 34b9e2c..183427a 100644 --- a/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -final class RSGroupInfoManagerImpl +final class RSGroupInfoManagerImpl extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements RSGroupInfoManager This is an implementation of RSGroupInfoManager which makes @@ -467,7 +467,7 @@ implements LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -476,7 +476,7 @@ implements RSGROUP_TABLE_DESC -private static finalHTableDescriptor RSGROUP_TABLE_DESC +private static finalHTableDescriptor RSGROUP_TABLE_DESC Table descriptor for hbase:rsgroup catalog table @@ -486,7 +486,7 @@ implements rsGroupMap -private volatilehttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,RSGroupInfo rsGroupMap +private volatilehttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,RSGroupInfo rsGroupMap @@ -495,7 +495,7 @@ implements tableMap -private volatilehttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapTableName,https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String tableMap +private volatilehttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapTableName,https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String tableMap @@ -504,7 +504,7 @@ implements masterServices -private finalMasterServices masterServices +private finalMasterServices masterServices @@ -513,7 +513,7 @@ implements rsGroupTable -privateTable rsGroupTable +privateTable rsGroupTable @@ -522,7 +522,7 @@ implements conn -private finalClusterConnection conn +private finalClusterConnection conn @@ -531,7 +531,7 @@ implements watcher -private finalZKWatcher watcher +private finalZKWatcher watcher @@ -540,7 +540,7 @@ implements rsGroupStartupWorker -private finalRSGroupInfoManagerImpl.RSGroupStartupWorker rsGroupStartupWorker +private finalRSGroupInfoManagerImpl.RSGroupStartupWorker rsGroupStartupWorker @@ -549,7 +549,7 @@ implements prevRSGroups -privatehttps://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Sethttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String prevRSGroups +privatehttps://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Sethttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String prevRSGroups @@ -558,7 +558,7 @@ implements serverEventsListenerThread -private finalRSGroupInfoManagerImpl.ServerEventsListenerThread serverEventsListenerThread +private finalRSGroupInfoManagerImpl.ServerEventsListenerThread serverEventsListenerThread @@ -567,7 +567,7 @@ implements failedOpenUpdaterThread -privateRSGroupInfoManagerImpl.FailedOpenUpdaterThread failedOpenUpdaterThread +privateRSGroupInfoManagerImpl.FailedOpenUpdaterThread failedOpenUpdaterThread @@ -584,7 +584,7 @@ implements RSGroupInfoManagerImpl -privateRSGroupInfoManagerImpl(MasterServicesmasterServices) +privateRSGroupInfoManagerImpl(MasterServicesmasterServices) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -606,7 +606,7 @@ implements init -privatevoidinit() +privatevoidinit() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -620,7 +620,7 @@ implements getInstance
[11/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.CompletedSnaphotDirectoriesFilter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.CompletedSnaphotDirectoriesFilter.html b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.CompletedSnaphotDirectoriesFilter.html index 63a00a7..32d6965 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.CompletedSnaphotDirectoriesFilter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.CompletedSnaphotDirectoriesFilter.html @@ -43,7 +43,7 @@ 035import org.apache.hadoop.hbase.security.User; 036import org.apache.hadoop.hbase.security.access.AccessControlLists; 037import org.apache.hadoop.hbase.security.access.ShadedAccessControlUtil; -038import org.apache.hadoop.hbase.security.access.TablePermission; +038import org.apache.hadoop.hbase.security.access.UserPermission; 039import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; 040import org.apache.hadoop.hbase.util.FSUtils; 041import org.apache.yetus.audience.InterfaceAudience; @@ -443,10 +443,10 @@ 435 436 private static SnapshotDescription writeAclToSnapshotDescription(SnapshotDescription snapshot, 437 Configuration conf) throws IOException { -438ListMultimapString, TablePermission perms = -439User.runAsLoginUser(new PrivilegedExceptionActionListMultimapString, TablePermission() { +438ListMultimapString, UserPermission perms = +439User.runAsLoginUser(new PrivilegedExceptionActionListMultimapString, UserPermission() { 440 @Override -441 public ListMultimapString, TablePermission run() throws Exception { +441 public ListMultimapString, UserPermission run() throws Exception { 442return AccessControlLists.getTablePermissions(conf, 443 TableName.valueOf(snapshot.getTable())); 444 } http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.html b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.html index 63a00a7..32d6965 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotDescriptionUtils.html @@ -43,7 +43,7 @@ 035import org.apache.hadoop.hbase.security.User; 036import org.apache.hadoop.hbase.security.access.AccessControlLists; 037import org.apache.hadoop.hbase.security.access.ShadedAccessControlUtil; -038import org.apache.hadoop.hbase.security.access.TablePermission; +038import org.apache.hadoop.hbase.security.access.UserPermission; 039import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; 040import org.apache.hadoop.hbase.util.FSUtils; 041import org.apache.yetus.audience.InterfaceAudience; @@ -443,10 +443,10 @@ 435 436 private static SnapshotDescription writeAclToSnapshotDescription(SnapshotDescription snapshot, 437 Configuration conf) throws IOException { -438ListMultimapString, TablePermission perms = -439User.runAsLoginUser(new PrivilegedExceptionActionListMultimapString, TablePermission() { +438ListMultimapString, UserPermission perms = +439User.runAsLoginUser(new PrivilegedExceptionActionListMultimapString, UserPermission() { 440 @Override -441 public ListMultimapString, TablePermission run() throws Exception { +441 public ListMultimapString, UserPermission run() throws Exception { 442return AccessControlLists.getTablePermissions(conf, 443 TableName.valueOf(snapshot.getTable())); 444 } http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/downloads.html -- diff --git a/downloads.html b/downloads.html index 5c0d243..405ed9e 100644 --- a/downloads.html +++ b/downloads.html @@ -7,7 +7,7 @@ - + Apache HBase Apache HBase Downloads @@ -461,7 +461,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-11-14 + Last Published: 2018-11-15 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/export_control.html -- diff --git a/export_control.html b/export_control.html index
[11/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/testdevapidocs/org/apache/hadoop/hbase/master/procedure/TestSchedulerQueueDeadLock.TableExclusiveProcedureWithId.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/master/procedure/TestSchedulerQueueDeadLock.TableExclusiveProcedureWithId.html b/testdevapidocs/org/apache/hadoop/hbase/master/procedure/TestSchedulerQueueDeadLock.TableExclusiveProcedureWithId.html new file mode 100644 index 000..0097e60 --- /dev/null +++ b/testdevapidocs/org/apache/hadoop/hbase/master/procedure/TestSchedulerQueueDeadLock.TableExclusiveProcedureWithId.html @@ -0,0 +1,357 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +TestSchedulerQueueDeadLock.TableExclusiveProcedureWithId (Apache HBase 3.0.0-SNAPSHOT Test API) + + + + + +var methods = {"i0":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.master.procedure +Class TestSchedulerQueueDeadLock.TableExclusiveProcedureWithId + + + +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.procedure2.ProcedureTEnv + + +org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility.NoopProcedureTestSchedulerQueueDeadLock.TestEnv + + +org.apache.hadoop.hbase.master.procedure.TestSchedulerQueueDeadLock.TableExclusiveProcedure + + +org.apache.hadoop.hbase.master.procedure.TestSchedulerQueueDeadLock.TableExclusiveProcedureWithId + + + + + + + + + + + + + +All Implemented Interfaces: +https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">Comparableorg.apache.hadoop.hbase.procedure2.ProcedureTestSchedulerQueueDeadLock.TestEnv, org.apache.hadoop.hbase.master.procedure.TableProcedureInterface + + +Enclosing class: +TestSchedulerQueueDeadLock + + + +public static final class TestSchedulerQueueDeadLock.TableExclusiveProcedureWithId +extends TestSchedulerQueueDeadLock.TableExclusiveProcedure + + + + + + + + + + + +Nested Class Summary + + + + +Nested classes/interfaces inherited from classorg.apache.hadoop.hbase.procedure2.Procedure +org.apache.hadoop.hbase.procedure2.Procedure.LockState + + + + + +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.master.procedure.TableProcedureInterface +org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType + + + + + + + + +Field Summary + + + + +Fields inherited from classorg.apache.hadoop.hbase.procedure2.Procedure +NO_PROC_ID, NO_TIMEOUT + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +TableExclusiveProcedureWithId() + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +protected void +setProcId(longprocId) + + + + + + +Methods inherited from classorg.apache.hadoop.hbase.master.procedure.TestSchedulerQueueDeadLock.TableExclusiveProcedure +acquireLock, execute, getTableName, getTableOperationType, holdLock, releaseLock + + + + + +Methods inherited from classorg.apache.hadoop.hbase.procedure2.ProcedureTestingUtility.NoopProcedure +abort, deserializeStateData, rollback, serializeStateData + + + + + +Methods inherited from classorg.apache.hadoop.hbase.procedure2.Procedure +addStackIndex, afterReplay, beforeReplay, bypass, compareTo, completionCleanup, doExecute, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isLockedWhenLoading, isRunnable, isSuccess,
[11/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/BitSetNode.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/BitSetNode.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/BitSetNode.html index be5c3fc..839d12b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/BitSetNode.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/BitSetNode.html @@ -139,294 +139,296 @@ 131 132 public BitSetNode(BitSetNode other, boolean resetDelete) { 133this.start = other.start; -134this.partial = other.partial; -135this.modified = other.modified.clone(); -136// The resetDelete will be set to true when building cleanup tracker. -137// The intention here is that, if a procedure is not modified in this tracker, then we do not -138// need to take care of it, so we will set deleted to true for these bits, i.e, if modified is -139// 0, then we set deleted to 1, otherwise keep it as is. So here, the equation is -140// deleted |= ~modified, i.e, -141if (resetDelete) { -142 this.deleted = new long[other.deleted.length]; -143 for (int i = 0; i this.deleted.length; ++i) { -144this.deleted[i] |= ~(other.modified[i]); -145 } -146} else { -147 this.deleted = other.deleted.clone(); -148} -149 } -150 -151 public void insertOrUpdate(final long procId) { -152updateState(procId, false); -153 } -154 -155 public void delete(final long procId) { -156updateState(procId, true); -157 } -158 -159 public long getStart() { -160return start; -161 } -162 -163 public long getEnd() { -164return start + (modified.length ADDRESS_BITS_PER_WORD) - 1; -165 } -166 -167 public boolean contains(final long procId) { -168return start = procId procId = getEnd(); -169 } -170 -171 public DeleteState isDeleted(final long procId) { -172int bitmapIndex = getBitmapIndex(procId); -173int wordIndex = bitmapIndex ADDRESS_BITS_PER_WORD; -174if (wordIndex = deleted.length) { -175 return DeleteState.MAYBE; -176} -177return (deleted[wordIndex] (1L bitmapIndex)) != 0 ? DeleteState.YES : DeleteState.NO; -178 } -179 -180 public boolean isModified(long procId) { -181int bitmapIndex = getBitmapIndex(procId); -182int wordIndex = bitmapIndex ADDRESS_BITS_PER_WORD; -183if (wordIndex = modified.length) { -184 return false; -185} -186return (modified[wordIndex] (1L bitmapIndex)) != 0; -187 } -188 -189 /** -190 * @return true, if all the procedures has been modified. -191 */ -192 public boolean isAllModified() { -193// TODO: cache the value -194for (int i = 0; i modified.length; ++i) { -195 if ((modified[i] | deleted[i]) != WORD_MASK) { -196return false; -197 } -198} -199return true; -200 } -201 -202 /** -203 * @return all the active procedure ids in this bit set. -204 */ -205 public long[] getActiveProcIds() { -206ListLong procIds = new ArrayList(); -207for (int wordIndex = 0; wordIndex modified.length; wordIndex++) { -208 if (deleted[wordIndex] == WORD_MASK || modified[wordIndex] == 0) { -209// This should be the common case, where most procedures has been deleted. -210continue; -211 } -212 long baseProcId = getStart() + (wordIndex ADDRESS_BITS_PER_WORD); -213 for (int i = 0; i (1 ADDRESS_BITS_PER_WORD); i++) { -214long mask = 1L i; -215if ((deleted[wordIndex] mask) == 0 (modified[wordIndex] mask) != 0) { -216 procIds.add(baseProcId + i); -217} -218 } -219} -220return procIds.stream().mapToLong(Long::longValue).toArray(); -221 } -222 -223 /** -224 * @return true, if there are no active procedures in this BitSetNode, else false. -225 */ -226 public boolean isEmpty() { -227// TODO: cache the value -228for (int i = 0; i deleted.length; ++i) { -229 if (deleted[i] != WORD_MASK) { -230return false; -231 } -232} -233return true; -234 } -235 -236 public void resetModified() { -237Arrays.fill(modified, 0); -238 } -239 -240 public void unsetPartialFlag() { -241partial = false; -242for (int i = 0; i modified.length; ++i) { -243 for (int j = 0; j BITS_PER_WORD; ++j) { -244if ((modified[i] (1L j)) == 0) { -245 deleted[i] |= (1L j); -246} -247 } -248} -249 } -250 -251 /** -252 * Convert to -253 * org.apache.hadoop.hbase.protobuf.generated.ProcedureProtos.ProcedureStoreTracker.TrackerNode -254 * protobuf. -255 */ -256 public ProcedureProtos.ProcedureStoreTracker.TrackerNode convert() { -257 ProcedureProtos.ProcedureStoreTracker.TrackerNode.Builder builder = -258
[11/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testdevapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/TestWALProcedureStore.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/TestWALProcedureStore.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/TestWALProcedureStore.html index 8c65786..3a1281c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/TestWALProcedureStore.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/TestWALProcedureStore.html @@ -28,976 +28,860 @@ 020import static org.junit.Assert.assertEquals; 021import static org.junit.Assert.assertFalse; 022import static org.junit.Assert.assertTrue; -023import static org.junit.Assert.fail; -024 -025import java.io.FileNotFoundException; -026import java.io.IOException; -027import java.io.InputStream; -028import java.io.OutputStream; -029import java.util.ArrayList; -030import java.util.Arrays; -031import java.util.Comparator; -032import java.util.HashSet; -033import java.util.Set; -034import org.apache.hadoop.conf.Configuration; -035import org.apache.hadoop.fs.FileStatus; -036import org.apache.hadoop.fs.FileSystem; -037import org.apache.hadoop.fs.Path; -038import org.apache.hadoop.hbase.HBaseClassTestRule; -039import org.apache.hadoop.hbase.HBaseCommonTestingUtility; -040import org.apache.hadoop.hbase.procedure2.Procedure; -041import org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer; -042import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility; -043import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility.LoadCounter; -044import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility.TestProcedure; -045import org.apache.hadoop.hbase.procedure2.SequentialProcedure; -046import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; -047import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; -048import org.apache.hadoop.hbase.procedure2.store.ProcedureStoreTracker; -049import org.apache.hadoop.hbase.testclassification.MasterTests; -050import org.apache.hadoop.hbase.testclassification.SmallTests; -051import org.apache.hadoop.io.IOUtils; -052import org.junit.After; -053import org.junit.Before; -054import org.junit.ClassRule; -055import org.junit.Test; -056import org.junit.experimental.categories.Category; -057import org.mockito.Mockito; -058import org.mockito.invocation.InvocationOnMock; -059import org.mockito.stubbing.Answer; -060import org.slf4j.Logger; -061import org.slf4j.LoggerFactory; +023 +024import java.io.FileNotFoundException; +025import java.io.IOException; +026import java.io.InputStream; +027import java.io.OutputStream; +028import java.util.ArrayList; +029import java.util.Arrays; +030import java.util.Comparator; +031import java.util.HashSet; +032import java.util.Set; +033import org.apache.hadoop.conf.Configuration; +034import org.apache.hadoop.fs.FileStatus; +035import org.apache.hadoop.fs.FileSystem; +036import org.apache.hadoop.fs.Path; +037import org.apache.hadoop.hbase.HBaseClassTestRule; +038import org.apache.hadoop.hbase.HBaseCommonTestingUtility; +039import org.apache.hadoop.hbase.procedure2.Procedure; +040import org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer; +041import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility; +042import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility.LoadCounter; +043import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility.TestProcedure; +044import org.apache.hadoop.hbase.procedure2.SequentialProcedure; +045import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; +046import org.apache.hadoop.hbase.procedure2.store.ProcedureStoreTracker; +047import org.apache.hadoop.hbase.testclassification.MasterTests; +048import org.apache.hadoop.hbase.testclassification.SmallTests; +049import org.apache.hadoop.io.IOUtils; +050import org.junit.After; +051import org.junit.Before; +052import org.junit.ClassRule; +053import org.junit.Test; +054import org.junit.experimental.categories.Category; +055import org.mockito.Mockito; +056import org.mockito.invocation.InvocationOnMock; +057import org.mockito.stubbing.Answer; +058import org.slf4j.Logger; +059import org.slf4j.LoggerFactory; +060 +061import org.apache.hbase.thirdparty.com.google.protobuf.Int64Value; 062 -063import org.apache.hbase.thirdparty.com.google.protobuf.Int64Value; -064 -065@Category({MasterTests.class, SmallTests.class}) -066public class TestWALProcedureStore { -067 -068 @ClassRule -069 public static final HBaseClassTestRule CLASS_RULE = -070 HBaseClassTestRule.forClass(TestWALProcedureStore.class); +063@Category({MasterTests.class, SmallTests.class}) +064public class TestWALProcedureStore { +065 +066 @ClassRule +067 public static final HBaseClassTestRule CLASS_RULE = +068
[11/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html b/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html index 773eb00..fa380ee 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html @@ -180,7 +180,7 @@ extends Procedure -abort, acquireLock, addStackIndex, afterReplay, beforeReplay, bypass, compareTo, completionCleanup, deserializeStateData, doAcquireLock, doExecute, doReleaseLock, doRollback, elapsedTime, execute, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutT imestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, lockedWhenLoading, needPersistence, releaseLock, removeStackIndex, resetPersistence, restoreLock, rollback, serializeStateData, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, s etOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, skipPersistence, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, tryRunnable, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted +abort, acquireLock, addStackIndex, afterReplay, beforeReplay, bypass, compareTo, completionCleanup, deserializeStateData, doAcquireLock, doExecute, doReleaseLock, doRollback, elapsedTime, execute, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#hasChildren--">hasChildren, > href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#hasException--">hasException, > href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#hasLock--">hasLock, > href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#hasOwner--">hasOwner, > href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#hasParent--">hasParent, > href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#hasTimeout--">hasTimeout, > href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#haveSameParent-org.apache.hadoop.hbase.procedure2.Procedure-org.apache.hadoop.hbase.procedure2.Procedure-">haveSameParent, > href="../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#holdLock-TEnvironment-">holdLock, > incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, lockedWhenLoading, needPersistence, releaseLock, removeStackIndex, resetPersistence, restoreLock, rollback, serializeStateData, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTime out, setTimeoutFailure, shouldWaitClientAck, skipPersistence, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, tryRunnable, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html index adb8b5d..dfd7844 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html @@ -652,22 +652,28 @@ Procedure?proc) +static https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException +MasterProcedureUtil.unwrapRemoteIOException(Procedureproc) +This is a version of
[11/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/procedure2/store/TestProcedureStoreTracker.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/procedure2/store/TestProcedureStoreTracker.html b/testdevapidocs/org/apache/hadoop/hbase/procedure2/store/TestProcedureStoreTracker.html index 5348cd0..6ba8c94 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/procedure2/store/TestProcedureStoreTracker.html +++ b/testdevapidocs/org/apache/hadoop/hbase/procedure2/store/TestProcedureStoreTracker.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab"; -PrevClass +PrevClass NextClass @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab"; -public class TestProcedureStoreTracker +public class TestProcedureStoreTracker extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -179,26 +179,30 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html void -testGetActiveProcIds() +testGetActiveMinProcId() void -testLoad() +testGetActiveProcIds() void -testPartialTracker() +testLoad() void -testRandLoad() +testPartialTracker() void -testSeqInsertAndDelete() +testRandLoad() void +testSeqInsertAndDelete() + + +void testSetDeletedIfModified() @@ -229,7 +233,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html CLASS_RULE -public static finalHBaseClassTestRule CLASS_RULE +public static finalHBaseClassTestRule CLASS_RULE @@ -238,7 +242,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -255,7 +259,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html TestProcedureStoreTracker -publicTestProcedureStoreTracker() +publicTestProcedureStoreTracker() @@ -272,7 +276,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html testSeqInsertAndDelete -publicvoidtestSeqInsertAndDelete() +publicvoidtestSeqInsertAndDelete() @@ -281,7 +285,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html testPartialTracker -publicvoidtestPartialTracker() +publicvoidtestPartialTracker() @@ -290,7 +294,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html testBasicCRUD -publicvoidtestBasicCRUD() +publicvoidtestBasicCRUD() @@ -299,7 +303,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html testRandLoad -publicvoidtestRandLoad() +publicvoidtestRandLoad() @@ -308,7 +312,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html testLoad -publicvoidtestLoad() +publicvoidtestLoad() @@ -317,7 +321,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html testDelete -publicvoidtestDelete() +publicvoidtestDelete() @@ -326,16 +330,25 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html testSetDeletedIfModified -publicvoidtestSetDeletedIfModified() +publicvoidtestSetDeletedIfModified() - + testGetActiveProcIds -publicvoidtestGetActiveProcIds() +publicvoidtestGetActiveProcIds() + + + + + + + +testGetActiveMinProcId +publicvoidtestGetActiveMinProcId() @@ -366,7 +379,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -PrevClass +PrevClass NextClass http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/procedure2/store/class-use/TestBitSetNode.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/procedure2/store/class-use/TestBitSetNode.html b/testdevapidocs/org/apache/hadoop/hbase/procedure2/store/class-use/TestBitSetNode.html new file mode 100644 index 000..1297c80 --- /dev/null +++ b/testdevapidocs/org/apache/hadoop/hbase/procedure2/store/class-use/TestBitSetNode.html @@ -0,0 +1,125 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +Uses of Class org.apache.hadoop.hbase.procedure2.store.TestBitSetNode (Apache HBase 3.0.0-SNAPSHOT Test API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase +Class StartMiniClusterOption.Builder + + + +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.StartMiniClusterOption.Builder + + + + + + + +Enclosing class: +StartMiniClusterOption + + + +public static final class StartMiniClusterOption.Builder +extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +Builder pattern for creating an StartMiniClusterOption. + + The default values of its fields should be considered public and constant. Changing the default + values may cause other tests fail. + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +private boolean +createRootDir + + +private boolean +createWALDir + + +private https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String[] +dataNodeHosts + + +private https://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends org.apache.hadoop.hbase.master.HMaster +masterClass + + +private int +numDataNodes + + +private int +numMasters + + +private int +numRegionServers + + +private int +numZkServers + + +private https://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends MiniHBaseCluster.MiniHBaseClusterRegionServer +rsClass + + +private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer +rsPorts + + + + + + + + + +Constructor Summary + +Constructors + +Modifier +Constructor and Description + + +private +Builder() + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +StartMiniClusterOption +build() + + +StartMiniClusterOption.Builder +createRootDir(booleancreateRootDir) + + +StartMiniClusterOption.Builder +createWALDir(booleancreateWALDir) + + +StartMiniClusterOption.Builder +dataNodeHosts(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String[]dataNodeHosts) + + +StartMiniClusterOption.Builder +masterClass(https://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends org.apache.hadoop.hbase.master.HMastermasterClass) + + +StartMiniClusterOption.Builder +numDataNodes(intnumDataNodes) + + +StartMiniClusterOption.Builder +numMasters(intnumMasters) + + +StartMiniClusterOption.Builder +numRegionServers(intnumRegionServers) + + +StartMiniClusterOption.Builder +numZkServers(intnumZkServers) + + +StartMiniClusterOption.Builder +rsClass(https://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends
[11/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/devapidocs/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html b/devapidocs/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html index dec3733..037ad83 100644 --- a/devapidocs/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html +++ b/devapidocs/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":9,"i3":10,"i4":10,"i5":42,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":42,"i14":10,"i15":10,"i16":10,"i17":9,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10}; +var methods = {"i0":10,"i1":10,"i2":9,"i3":10,"i4":10,"i5":10,"i6":42,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":42,"i15":10,"i16":10,"i17":10,"i18":10,"i19":9,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -124,7 +124,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Public -public class SingleColumnValueFilter +public class SingleColumnValueFilter extends FilterBase This filter is used to filter cells based on value. It takes a CompareFilter.CompareOp operator (equal, greater, not equal, etc), and either a byte [] value or @@ -337,22 +337,26 @@ extends createFilterFromArguments(https://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true; title="class or interface in java.util">ArrayListbyte[]filterArguments) +boolean +equals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectobj) + + Filter.ReturnCode filterCell(Cellc) A way to filter based on the column family, column qualifier and/or the column value. - + private boolean filterColumnValue(Cellcell) - + Filter.ReturnCode filterKeyValue(Cellc) Deprecated. - + boolean filterRow() Filters that never filter by rows based on previously gathered state from @@ -360,37 +364,37 @@ extends - + boolean filterRowKey(Cellcell) Filters a row based on the row key. - + ByteArrayComparable getComparator() - + CompareOperator getCompareOperator() - + byte[] getFamily() - + boolean getFilterIfMissing() Get whether entire row should be filtered if column is not found. - + boolean getLatestVersionOnly() Get whether only the latest version of the column value should be compared. - + CompareFilter.CompareOp getOperator() Deprecated. @@ -398,53 +402,57 @@ extends - + byte[] getQualifier() - + boolean hasFilterRow() Fitlers that never filter by modifying the returned List of Cells can inherit this implementation that does nothing. - + +int +hashCode() + + boolean isFamilyEssential(byte[]name) The only CF this filter needs is given column family. - + static SingleColumnValueFilter parseFrom(byte[]pbBytes) - + void reset() Filters that are purely stateless and do nothing in their reset() methods can inherit this null/empty implementation. - + void setFilterIfMissing(booleanfilterIfMissing) Set whether entire row should be filtered if column is not found. - + void setLatestVersionOnly(booleanlatestVersionOnly) Set whether only the latest version of the column value should be compared. - + byte[] toByteArray() Return length 0 byte array for Filters that don't require special serialization - + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String toString() Return filter's info for debugging and logging purpose. @@ -470,7 +478,7 @@ extends Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object -https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
[11/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html index 2709ea3..4a11f27 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.html @@ -37,309 +37,299 @@ 029import org.apache.hadoop.hbase.executor.EventType; 030import org.apache.hadoop.hbase.regionserver.HRegion; 031import org.apache.hadoop.hbase.regionserver.Region; -032import org.apache.hadoop.hbase.regionserver.RegionServerAccounting; -033import org.apache.hadoop.hbase.regionserver.RegionServerServices; -034import org.apache.hadoop.hbase.regionserver.RegionServerServices.PostOpenDeployContext; -035import org.apache.hadoop.hbase.regionserver.RegionServerServices.RegionStateTransitionContext; -036import org.apache.hadoop.hbase.util.CancelableProgressable; -037import org.apache.yetus.audience.InterfaceAudience; -038import org.slf4j.Logger; -039import org.slf4j.LoggerFactory; -040import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode; -041/** -042 * Handles opening of a region on a region server. -043 * p -044 * This is executed after receiving an OPEN RPC from the master or client. -045 */ -046@InterfaceAudience.Private -047public class OpenRegionHandler extends EventHandler { -048 private static final Logger LOG = LoggerFactory.getLogger(OpenRegionHandler.class); -049 -050 protected final RegionServerServices rsServices; -051 -052 private final RegionInfo regionInfo; -053 private final TableDescriptor htd; -054 private final long masterSystemTime; -055 -056 public OpenRegionHandler(final Server server, -057 final RegionServerServices rsServices, RegionInfo regionInfo, -058 TableDescriptor htd, long masterSystemTime) { -059this(server, rsServices, regionInfo, htd, masterSystemTime, EventType.M_RS_OPEN_REGION); -060 } -061 -062 protected OpenRegionHandler(final Server server, -063 final RegionServerServices rsServices, final RegionInfo regionInfo, -064 final TableDescriptor htd, long masterSystemTime, EventType eventType) { -065super(server, eventType); -066this.rsServices = rsServices; -067this.regionInfo = regionInfo; -068this.htd = htd; -069this.masterSystemTime = masterSystemTime; -070 } -071 -072 public RegionInfo getRegionInfo() { -073return regionInfo; -074 } -075 -076 @Override -077 public void process() throws IOException { -078boolean openSuccessful = false; -079final String regionName = regionInfo.getRegionNameAsString(); -080HRegion region = null; -081 -082try { -083 if (this.server.isStopped() || this.rsServices.isStopping()) { -084return; -085 } -086 final String encodedName = regionInfo.getEncodedName(); -087 -088 // 2 different difficult situations can occur -089 // 1) The opening was cancelled. This is an expected situation -090 // 2) The region is now marked as online while we're suppose to open. This would be a bug. -091 -092 // Check that this region is not already online -093 if (this.rsServices.getRegion(encodedName) != null) { -094LOG.error("Region " + encodedName + -095" was already online when we started processing the opening. " + -096"Marking this new attempt as failed"); -097return; -098 } -099 -100 // Check that we're still supposed to open the region. -101 // If fails, just return. Someone stole the region from under us. -102 if (!isRegionStillOpening()){ -103LOG.error("Region " + encodedName + " opening cancelled"); -104return; -105 } -106 -107 // Open region. After a successful open, failures in subsequent -108 // processing needs to do a close as part of cleanup. -109 region = openRegion(); -110 if (region == null) { -111return; -112 } -113 -114 if (!updateMeta(region, masterSystemTime) || this.server.isStopped() || -115 this.rsServices.isStopping()) { -116return; -117 } -118 -119 if (!isRegionStillOpening()) { -120return; -121 } -122 -123 // Successful region open, and add it to MutableOnlineRegions -124 this.rsServices.addRegion(region); -125 openSuccessful = true; -126 -127 // Done! Successful region open -128 LOG.debug("Opened " + regionName + " on " + this.server.getServerName()); -129} finally { -130 // Do all clean up here -131 if
[11/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html index cfe74f4..7d4b201 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html @@ -153,11 +153,18 @@ +private void +HMaster.checkUnsupportedProcedure(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends Procedure,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListProcedureMasterProcedureEnvprocsByType) +In HBASE-20811, we have introduced a new TRSP to assign/unassign/move regions, and it is + incompatible with the old AssignProcedure/UnassignProcedure/MoveRegionProcedure. + + + void MasterCoprocessorHost.preAbortProcedure(ProcedureExecutorMasterProcedureEnvprocEnv, longprocId) - + private long ClusterSchemaServiceImpl.submitProcedure(ProcedureMasterProcedureEnvprocedure, NonceKeynonceKey) @@ -191,45 +198,58 @@ boolean -MoveRegionProcedure.abort(MasterProcedureEnvenv) +MoveRegionProcedure.abort(MasterProcedureEnvenv) +Deprecated. + protected boolean -SplitTableRegionProcedure.abort(MasterProcedureEnvenv) +RegionRemoteProcedureBase.abort(MasterProcedureEnvenv) protected boolean -MergeTableRegionsProcedure.abort(MasterProcedureEnvenv) +SplitTableRegionProcedure.abort(MasterProcedureEnvenv) protected boolean -RegionTransitionProcedure.abort(MasterProcedureEnvenv) +MergeTableRegionsProcedure.abort(MasterProcedureEnvenv) +protected boolean +RegionTransitionProcedure.abort(MasterProcedureEnvenv) +Deprecated. + + + protected Procedure.LockState MergeTableRegionsProcedure.acquireLock(MasterProcedureEnvenv) + +static TransitRegionStateProcedure +TransitRegionStateProcedure.assign(MasterProcedureEnvenv, + RegionInforegion, + ServerNametargetServer) + -protected Procedure.LockState -RegionTransitionProcedure.acquireLock(MasterProcedureEnvenv) +(package private) static void +AssignmentManagerUtil.checkClosedRegion(MasterProcedureEnvenv, + RegionInforegionInfo) -protected boolean -RegionTransitionProcedure.addToRemoteDispatcher(MasterProcedureEnvenv, - ServerNametargetServer) -Be careful! At the end of this method, the procedure has either succeeded - and this procedure has been set into a suspended state OR, we failed and - this procedure has been put back on the scheduler ready for another worker - to pick it up. - +private void +SplitTableRegionProcedure.checkClosedRegions(MasterProcedureEnvenv) +private void +MergeTableRegionsProcedure.checkClosedRegions(MasterProcedureEnvenv) + + private static void MergeTableRegionsProcedure.checkRegionsToMerge(MasterProcedureEnvenv, RegionInfo[]regionsToMerge, booleanforcible) - + private static void MergeTableRegionsProcedure.checkRegionsToMerge(MasterProcedureEnvenv, RegionInforegionToMergeA, @@ -238,7 +258,7 @@ One time checks. - + private void SplitTableRegionProcedure.checkSplittable(MasterProcedureEnvenv, RegionInforegionToSplit, @@ -246,21 +266,48 @@ Check whether the region is splittable - + private void MergeTableRegionsProcedure.cleanupMergedRegion(MasterProcedureEnvenv) Clean up a merged region + +private void +TransitRegionStateProcedure.closeRegion(MasterProcedureEnvenv, + RegionStateNoderegionNode) + + +private StateMachineProcedure.Flow +TransitRegionStateProcedure.confirmClosed(MasterProcedureEnvenv, + RegionStateNoderegionNode) + + +private StateMachineProcedure.Flow +TransitRegionStateProcedure.confirmOpened(MasterProcedureEnvenv, + RegionStateNoderegionNode) + + +private TransitRegionStateProcedure[] +SplitTableRegionProcedure.createAssignProcedures(MasterProcedureEnvenv) + + +private TransitRegionStateProcedure[] +MergeTableRegionsProcedure.createAssignProcedures(MasterProcedureEnvenv) + -private AssignProcedure[] -SplitTableRegionProcedure.createAssignProcedures(MasterProcedureEnvenv, - intregionReplication) +private static TransitRegionStateProcedure[] +AssignmentManagerUtil.createAssignProcedures(MasterProcedureEnvenv, + https://docs.oracle.com/javase/8/docs/api/java/util/stream/Stream.html?is-external=true; title="class or interface in
[11/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.OutputSink.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.OutputSink.html b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.OutputSink.html index 63e4b46..514f830 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.OutputSink.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.OutputSink.html @@ -468,15 +468,15 @@ 460 * creating it if necessary. 461 * @param logEntry 462 * @param fileNameBeingSplit the file being split currently. Used to generate tmp file name. -463 * @param conf -464 * @return Path to file into which to dump split log edits. -465 * @throws IOException -466 */ -467 @SuppressWarnings("deprecation") -468 @VisibleForTesting -469 static Path getRegionSplitEditsPath(final Entry logEntry, String fileNameBeingSplit, -470 Configuration conf) -471 throws IOException { +463 * @param tmpDirName of the directory used to sideline old recovered edits file +464 * @param conf +465 * @return Path to file into which to dump split log edits. +466 * @throws IOException +467 */ +468 @SuppressWarnings("deprecation") +469 @VisibleForTesting +470 static Path getRegionSplitEditsPath(final Entry logEntry, String fileNameBeingSplit, +471 String tmpDirName, Configuration conf) throws IOException { 472FileSystem fs = FileSystem.get(conf); 473Path rootDir = FSUtils.getRootDir(conf); 474Path tableDir = FSUtils.getTableDir(rootDir, logEntry.getKey().getTableName()); @@ -491,7 +491,7 @@ 483 return null; 484} 485if (fs.exists(dir) fs.isFile(dir)) { -486 Path tmp = new Path("/tmp"); +486 Path tmp = new Path(tmpDirName); 487 if (!fs.exists(tmp)) { 488fs.mkdirs(tmp); 489 } @@ -1520,411 +1520,413 @@ 1512 * @return a path with a write for that path. caller should close. 1513 */ 1514WriterAndPath createWAP(byte[] region, Entry entry) throws IOException { -1515 Path regionedits = getRegionSplitEditsPath(entry, -1516 fileBeingSplit.getPath().getName(), conf); -1517 if (regionedits == null) { -1518return null; -1519 } -1520 FileSystem rootFs = FileSystem.get(conf); -1521 if (rootFs.exists(regionedits)) { -1522LOG.warn("Found old edits file. It could be the " -1523+ "result of a previous failed split attempt. Deleting " + regionedits + ", length=" -1524+ rootFs.getFileStatus(regionedits).getLen()); -1525if (!rootFs.delete(regionedits, false)) { -1526 LOG.warn("Failed delete of old {}", regionedits); -1527} -1528 } -1529 Writer w = createWriter(regionedits); -1530 LOG.debug("Creating writer path={}", regionedits); -1531 return new WriterAndPath(regionedits, w, entry.getKey().getSequenceId()); -1532} -1533 -1534void filterCellByStore(Entry logEntry) { -1535 Mapbyte[], Long maxSeqIdInStores = -1536 regionMaxSeqIdInStores.get(Bytes.toString(logEntry.getKey().getEncodedRegionName())); -1537 if (MapUtils.isEmpty(maxSeqIdInStores)) { -1538return; -1539 } -1540 // Create the array list for the cells that aren't filtered. -1541 // We make the assumption that most cells will be kept. -1542 ArrayListCell keptCells = new ArrayList(logEntry.getEdit().getCells().size()); -1543 for (Cell cell : logEntry.getEdit().getCells()) { -1544if (CellUtil.matchingFamily(cell, WALEdit.METAFAMILY)) { -1545 keptCells.add(cell); -1546} else { -1547 byte[] family = CellUtil.cloneFamily(cell); -1548 Long maxSeqId = maxSeqIdInStores.get(family); -1549 // Do not skip cell even if maxSeqId is null. Maybe we are in a rolling upgrade, -1550 // or the master was crashed before and we can not get the information. -1551 if (maxSeqId == null || maxSeqId.longValue() logEntry.getKey().getSequenceId()) { -1552keptCells.add(cell); -1553 } -1554} -1555 } -1556 -1557 // Anything in the keptCells array list is still live. -1558 // So rather than removing the cells from the array list -1559 // which would be an O(n^2) operation, we just replace the list -1560 logEntry.getEdit().setCells(keptCells); -1561} -1562 -1563@Override -1564public void append(RegionEntryBuffer buffer) throws IOException { -1565 appendBuffer(buffer, true); -1566} -1567 -1568WriterAndPath appendBuffer(RegionEntryBuffer buffer, boolean reusable) throws IOException{ -1569 ListEntry entries = buffer.entryBuffer; -1570 if (entries.isEmpty()) { -1571LOG.warn("got an empty buffer, skipping"); -1572return null; -1573 }
[11/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html index f2fd195..b293714 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html @@ -619,1696 +619,1698 @@ 611try { 612 long procId = 613 master.createTable(tableDescriptor, splitKeys, req.getNonceGroup(), req.getNonce()); -614 return CreateTableResponse.newBuilder().setProcId(procId).build(); -615} catch (IOException ioe) { -616 throw new ServiceException(ioe); -617} -618 } -619 -620 @Override -621 public DeleteColumnResponse deleteColumn(RpcController controller, -622 DeleteColumnRequest req) throws ServiceException { -623try { -624 long procId = master.deleteColumn( -625 ProtobufUtil.toTableName(req.getTableName()), -626 req.getColumnName().toByteArray(), -627req.getNonceGroup(), -628req.getNonce()); -629 if (procId == -1) { -630// This mean operation was not performed in server, so do not set any procId -631return DeleteColumnResponse.newBuilder().build(); -632 } else { -633return DeleteColumnResponse.newBuilder().setProcId(procId).build(); -634 } -635} catch (IOException ioe) { -636 throw new ServiceException(ioe); -637} -638 } -639 -640 @Override -641 public DeleteNamespaceResponse deleteNamespace(RpcController controller, -642 DeleteNamespaceRequest request) throws ServiceException { -643try { -644 long procId = master.deleteNamespace( -645request.getNamespaceName(), -646request.getNonceGroup(), -647request.getNonce()); -648 return DeleteNamespaceResponse.newBuilder().setProcId(procId).build(); -649} catch (IOException e) { -650 throw new ServiceException(e); -651} -652 } -653 -654 /** -655 * Execute Delete Snapshot operation. -656 * @return DeleteSnapshotResponse (a protobuf wrapped void) if the snapshot existed and was -657 *deleted properly. -658 * @throws ServiceException wrapping SnapshotDoesNotExistException if specified snapshot did not -659 *exist. -660 */ -661 @Override -662 public DeleteSnapshotResponse deleteSnapshot(RpcController controller, -663 DeleteSnapshotRequest request) throws ServiceException { -664try { -665 master.checkInitialized(); -666 master.snapshotManager.checkSnapshotSupport(); -667 -668 LOG.info(master.getClientIdAuditPrefix() + " delete " + request.getSnapshot()); -669 master.snapshotManager.deleteSnapshot(request.getSnapshot()); -670 return DeleteSnapshotResponse.newBuilder().build(); -671} catch (IOException e) { -672 throw new ServiceException(e); -673} -674 } -675 -676 @Override -677 public DeleteTableResponse deleteTable(RpcController controller, -678 DeleteTableRequest request) throws ServiceException { -679try { -680 long procId = master.deleteTable(ProtobufUtil.toTableName( -681 request.getTableName()), request.getNonceGroup(), request.getNonce()); -682 return DeleteTableResponse.newBuilder().setProcId(procId).build(); -683} catch (IOException ioe) { -684 throw new ServiceException(ioe); -685} -686 } -687 -688 @Override -689 public TruncateTableResponse truncateTable(RpcController controller, TruncateTableRequest request) -690 throws ServiceException { -691try { -692 long procId = master.truncateTable( -693 ProtobufUtil.toTableName(request.getTableName()), -694request.getPreserveSplits(), -695request.getNonceGroup(), -696request.getNonce()); -697 return TruncateTableResponse.newBuilder().setProcId(procId).build(); -698} catch (IOException ioe) { -699 throw new ServiceException(ioe); -700} -701 } -702 -703 @Override -704 public DisableTableResponse disableTable(RpcController controller, -705 DisableTableRequest request) throws ServiceException { -706try { -707 long procId = master.disableTable( -708 ProtobufUtil.toTableName(request.getTableName()), -709request.getNonceGroup(), -710request.getNonce()); -711 return DisableTableResponse.newBuilder().setProcId(procId).build(); -712} catch (IOException ioe) { -713 throw new ServiceException(ioe); -714} -715 } -716 -717 @Override -718 public EnableCatalogJanitorResponse enableCatalogJanitor(RpcController c, -719 EnableCatalogJanitorRequest req) throws ServiceException { -720 rpcPreCheck("enableCatalogJanitor"); -721return
[11/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcDuplexHandler.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcDuplexHandler.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcDuplexHandler.html index c5acc87..638f1aa 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcDuplexHandler.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcDuplexHandler.html @@ -25,239 +25,240 @@ 017 */ 018package org.apache.hadoop.hbase.ipc; 019 -020import org.apache.hbase.thirdparty.com.google.protobuf.Message; -021import org.apache.hbase.thirdparty.com.google.protobuf.Message.Builder; -022import org.apache.hbase.thirdparty.com.google.protobuf.TextFormat; -023 -024import org.apache.hbase.thirdparty.io.netty.buffer.ByteBuf; -025import org.apache.hbase.thirdparty.io.netty.buffer.ByteBufInputStream; -026import org.apache.hbase.thirdparty.io.netty.buffer.ByteBufOutputStream; -027import org.apache.hbase.thirdparty.io.netty.channel.ChannelDuplexHandler; -028import org.apache.hbase.thirdparty.io.netty.channel.ChannelHandlerContext; -029import org.apache.hbase.thirdparty.io.netty.channel.ChannelPromise; -030import org.apache.hbase.thirdparty.io.netty.handler.timeout.IdleStateEvent; -031import org.apache.hbase.thirdparty.io.netty.util.concurrent.PromiseCombiner; -032 -033import java.io.IOException; -034import java.util.HashMap; -035import java.util.Map; -036 -037import org.apache.hadoop.hbase.CellScanner; -038import org.apache.yetus.audience.InterfaceAudience; -039import org.slf4j.Logger; -040import org.slf4j.LoggerFactory; -041import org.apache.hadoop.hbase.codec.Codec; -042import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta; -043import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ExceptionResponse; -044import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -045import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ResponseHeader; -046import org.apache.hadoop.io.compress.CompressionCodec; -047import org.apache.hadoop.ipc.RemoteException; -048 -049/** -050 * The netty rpc handler. -051 * @since 2.0.0 -052 */ -053@InterfaceAudience.Private -054class NettyRpcDuplexHandler extends ChannelDuplexHandler { -055 -056 private static final Logger LOG = LoggerFactory.getLogger(NettyRpcDuplexHandler.class); -057 -058 private final NettyRpcConnection conn; -059 -060 private final CellBlockBuilder cellBlockBuilder; -061 -062 private final Codec codec; -063 -064 private final CompressionCodec compressor; -065 -066 private final MapInteger, Call id2Call = new HashMap(); -067 -068 public NettyRpcDuplexHandler(NettyRpcConnection conn, CellBlockBuilder cellBlockBuilder, -069 Codec codec, CompressionCodec compressor) { -070this.conn = conn; -071this.cellBlockBuilder = cellBlockBuilder; -072this.codec = codec; -073this.compressor = compressor; -074 -075 } -076 -077 private void writeRequest(ChannelHandlerContext ctx, Call call, ChannelPromise promise) -078 throws IOException { -079id2Call.put(call.id, call); -080ByteBuf cellBlock = cellBlockBuilder.buildCellBlock(codec, compressor, call.cells, ctx.alloc()); -081CellBlockMeta cellBlockMeta; -082if (cellBlock != null) { -083 CellBlockMeta.Builder cellBlockMetaBuilder = CellBlockMeta.newBuilder(); -084 cellBlockMetaBuilder.setLength(cellBlock.writerIndex()); -085 cellBlockMeta = cellBlockMetaBuilder.build(); -086} else { -087 cellBlockMeta = null; -088} -089RequestHeader requestHeader = IPCUtil.buildRequestHeader(call, cellBlockMeta); -090int sizeWithoutCellBlock = IPCUtil.getTotalSizeWhenWrittenDelimited(requestHeader, call.param); -091int totalSize = cellBlock != null ? sizeWithoutCellBlock + cellBlock.writerIndex() -092: sizeWithoutCellBlock; -093ByteBuf buf = ctx.alloc().buffer(sizeWithoutCellBlock + 4); -094buf.writeInt(totalSize); -095try (ByteBufOutputStream bbos = new ByteBufOutputStream(buf)) { -096 requestHeader.writeDelimitedTo(bbos); -097 if (call.param != null) { -098 call.param.writeDelimitedTo(bbos); -099 } -100 if (cellBlock != null) { -101ChannelPromise withoutCellBlockPromise = ctx.newPromise(); -102ctx.write(buf, withoutCellBlockPromise); -103ChannelPromise cellBlockPromise = ctx.newPromise(); -104ctx.write(cellBlock, cellBlockPromise); -105PromiseCombiner combiner = new PromiseCombiner(); -106 combiner.addAll(withoutCellBlockPromise, cellBlockPromise); -107combiner.finish(promise); -108 } else { -109ctx.write(buf, promise); -110 } -111} -112 } -113 -114 @Override -115 public void write(ChannelHandlerContext ctx, Object msg,
[11/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaRackCostFunction.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaRackCostFunction.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaRackCostFunction.html index 233dba3..91b9055 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaRackCostFunction.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaRackCostFunction.html @@ -540,1205 +540,1204 @@ 532 sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) - { 533DequeBalancerRegionLoad rLoads = oldLoads.get(Bytes.toString(regionName)); 534if (rLoads == null) { -535 // There was nothing there -536 rLoads = new ArrayDeque(); -537} else if (rLoads.size() = numRegionLoadsToRemember) { -538 rLoads.remove(); -539} -540rLoads.add(new BalancerRegionLoad(rm)); -541 loads.put(Bytes.toString(regionName), rLoads); -542 }); -543}); -544 -545for(CostFromRegionLoadFunction cost : regionLoadFunctions) { -546 cost.setLoads(loads); -547} -548 } -549 -550 protected void initCosts(Cluster cluster) { -551for (CostFunction c:costFunctions) { -552 c.init(cluster); -553} -554 } -555 -556 protected void updateCostsWithAction(Cluster cluster, Action action) { -557for (CostFunction c : costFunctions) { -558 c.postAction(action); -559} -560 } -561 -562 /** -563 * Get the names of the cost functions -564 */ -565 public String[] getCostFunctionNames() { -566if (costFunctions == null) return null; -567String[] ret = new String[costFunctions.length]; -568for (int i = 0; i costFunctions.length; i++) { -569 CostFunction c = costFunctions[i]; -570 ret[i] = c.getClass().getSimpleName(); -571} -572 -573return ret; -574 } -575 -576 /** -577 * This is the main cost function. It will compute a cost associated with a proposed cluster -578 * state. All different costs will be combined with their multipliers to produce a double cost. -579 * -580 * @param cluster The state of the cluster -581 * @param previousCost the previous cost. This is used as an early out. -582 * @return a double of a cost associated with the proposed cluster state. This cost is an -583 * aggregate of all individual cost functions. -584 */ -585 protected double computeCost(Cluster cluster, double previousCost) { -586double total = 0; -587 -588for (int i = 0; i costFunctions.length; i++) { -589 CostFunction c = costFunctions[i]; -590 this.tempFunctionCosts[i] = 0.0; -591 -592 if (c.getMultiplier() = 0) { -593continue; -594 } -595 -596 Float multiplier = c.getMultiplier(); -597 Double cost = c.cost(); -598 -599 this.tempFunctionCosts[i] = multiplier*cost; -600 total += this.tempFunctionCosts[i]; -601 -602 if (total previousCost) { -603break; -604 } -605} -606 -607return total; -608 } -609 -610 /** Generates a candidate action to be applied to the cluster for cost function search */ -611 abstract static class CandidateGenerator { -612abstract Cluster.Action generate(Cluster cluster); -613 -614/** -615 * From a list of regions pick a random one. Null can be returned which -616 * {@link StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region move -617 * rather than swap. -618 * -619 * @param clusterThe state of the cluster -620 * @param server index of the server -621 * @param chanceOfNoSwap Chance that this will decide to try a move rather -622 * than a swap. -623 * @return a random {@link RegionInfo} or null if an asymmetrical move is -624 * suggested. -625 */ -626protected int pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) { -627 // Check to see if this is just a move. -628 if (cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat() chanceOfNoSwap) { -629// signal a move only. -630return -1; -631 } -632 int rand = RANDOM.nextInt(cluster.regionsPerServer[server].length); -633 return cluster.regionsPerServer[server][rand]; -634 -635} -636protected int pickRandomServer(Cluster cluster) { -637 if (cluster.numServers 1) { -638return -1; -639 } -640 -641 return RANDOM.nextInt(cluster.numServers); -642} -643 -644protected int pickRandomRack(Cluster cluster) { -645 if (cluster.numRacks 1) {
[11/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/class-use/CacheEvictionStats.html -- diff --git a/apidocs/org/apache/hadoop/hbase/class-use/CacheEvictionStats.html b/apidocs/org/apache/hadoop/hbase/class-use/CacheEvictionStats.html index 08f91e2..166d35e 100644 --- a/apidocs/org/apache/hadoop/hbase/class-use/CacheEvictionStats.html +++ b/apidocs/org/apache/hadoop/hbase/class-use/CacheEvictionStats.html @@ -1,10 +1,10 @@ http://www.w3.org/TR/html4/loose.dtd;> - + -ç±» org.apache.hadoop.hbase.CacheEvictionStatsçä½¿ç¨ (Apache HBase 3.0.0-SNAPSHOT API) +Uses of Class org.apache.hadoop.hbase.CacheEvictionStats (Apache HBase 3.0.0-SNAPSHOT API) @@ -12,7 +12,7 @@ -æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã +JavaScript is disabled on your browser. -è·³è¿å¯¼èªé¾æ¥ +Skip navigation links - -æ¦è§ -ç¨åºå -ç±» -ä½¿ç¨ -æ -å·²è¿æ¶ -ç´¢å¼ -å¸®å© + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help -ä¸ä¸ä¸ª -ä¸ä¸ä¸ª +Prev +Next -æ¡æ¶ -æ æ¡æ¶ +Frames +NoFrames -ææç±» +AllClasses-ç±»ç使ç¨
+
org.apache.hadoop.hbase.CacheEvictionStatsUses of Class
org.apache.hadoop.hbase.CacheEvictionStats
- -