[31/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/thrift2/client/ThriftConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift2/client/ThriftConnection.html b/devapidocs/org/apache/hadoop/hbase/thrift2/client/ThriftConnection.html new file mode 100644 index 000..14563b5 --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/thrift2/client/ThriftConnection.html @@ -0,0 +1,881 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +ThriftConnection (Apache HBase 3.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.thrift2.client +Class ThriftConnection + + + +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.thrift2.client.ThriftConnection + + + + + + + +All Implemented Interfaces: +https://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable, https://docs.oracle.com/javase/8/docs/api/java/lang/AutoCloseable.html?is-external=true; title="class or interface in java.lang">AutoCloseable, Abortable, Connection + + + +@InterfaceAudience.Private +public class ThriftConnection +extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +implements Connection + + + + + + + + + + + +Nested Class Summary + +Nested Classes + +Modifier and Type +Class and Description + + +static class +ThriftConnection.DefaultThriftClientBuilder +the default thrift client builder. + + + +static class +ThriftConnection.HTTPThriftClientBuilder +the default thrift http client builder. + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +private ThriftClientBuilder +clientBuilder + + +private org.apache.hadoop.conf.Configuration +conf + + +private int +connectTimeout + + +private https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +host + + +private org.apache.http.client.HttpClient +httpClient + + +private boolean +httpClientCreated + + +private boolean +isClosed + + +private boolean +isCompact + + +private boolean +isFramed + + +private int +operationTimeout + + +private int +port + + +private User +user + + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +ThriftConnection(org.apache.hadoop.conf.Configurationconf, +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorServicepool, +Useruser) + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +void +abort(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringwhy, + https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablee) +Abort the server or client. + + + +void +close() + + +Admin +getAdmin() +Get a ThriftAdmin, ThriftAdmin is NOT thread safe + + + +BufferedMutator +getBufferedMutator(BufferedMutatorParamsparams) +Retrieve a BufferedMutator for performing client-side buffering of writes. + + + +BufferedMutator +getBufferedMutator(TableNametableName) + + Retrieve a BufferedMutator for performing client-side buffering of writes. + + + +ThriftClientBuilder +getClientBuilder() + + +org.apache.hadoop.conf.Configuration +getConfiguration() + + +int +getConnectTimeout() + +
[31/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html index cd65fc3..2139972 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html +++ b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class ShortCircuitMasterConnection +public class ShortCircuitMasterConnection extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements MasterKeepAliveConnection A short-circuit connection that can bypass the RPC layer (serialization, deserialization, @@ -398,161 +398,171 @@ implements +org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse +isRpcThrottleEnabled(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequestrequest) + + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse isSnapshotDone(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse isSplitOrMergeEnabled(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse listDecommissionedRegionServers(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse listNamespaceDescriptors(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse listReplicationPeers(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse listTableDescriptorsByNamespace(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequestrequest) - +
[31/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html index 243c9e4..19b5553 100644 --- a/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/package-tree.html @@ -293,9 +293,9 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.mapreduce.RowCounter.RowCounterMapper.Counters org.apache.hadoop.hbase.mapreduce.CellCounter.CellCounterMapper.Counters org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper.Counter +org.apache.hadoop.hbase.mapreduce.RowCounter.RowCounterMapper.Counters org.apache.hadoop.hbase.mapreduce.TableSplit.Version http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html index c49c655..848d41f 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html @@ -198,8 +198,8 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.LocalityType org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type +org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.LocalityType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/org/apache/hadoop/hbase/master/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html index 625c97b..bff0e70 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html @@ -347,10 +347,10 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage -org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus org.apache.hadoop.hbase.master.RegionState.State org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective +org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus +org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html index 5fd1c3a..653e33c 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html @@ -216,9 +216,9 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or
[31/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftServer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftServer.html b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftServer.html index c13d3bf..a59130a 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftServer.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftServer.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":9,"i2":10,"i3":10,"i4":10,"i5":9,"i6":10,"i7":10,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":10,"i19":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":9,"i4":10,"i5":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -103,12 +103,17 @@ var activeTableTab = "activeTableTab"; org.apache.hadoop.conf.Configured +org.apache.hadoop.hbase.thrift.ThriftServer + + org.apache.hadoop.hbase.thrift2.ThriftServer + + @@ -119,9 +124,8 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.LimitedPrivate(value="Tools") -public class ThriftServer -extends org.apache.hadoop.conf.Configured -implements org.apache.hadoop.util.Tool +public class ThriftServer +extends ThriftServer ThriftServer - this class starts up a Thrift server which implements the HBase API specified in the HbaseClient.thrift IDL file. @@ -143,39 +147,17 @@ implements org.apache.hadoop.util.Tool Field and Description -(package private) static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String -BACKLOG_CONF_KEY - - -static int -DEFAULT_LISTEN_PORT - - private static org.slf4j.Logger log - -private static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String -READ_TIMEOUT_OPTION - - -(package private) static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String -THRIFT_QOP_KEY -Thrift quality of protection configuration key. - - - -static int -THRIFT_SERVER_SOCKET_READ_TIMEOUT_DEFAULT - - -static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String -THRIFT_SERVER_SOCKET_READ_TIMEOUT_KEY -Amount of time in milliseconds before a server thread will timeout - waiting for client to send data on a connected socket. - - + + + + +Fields inherited from classorg.apache.hadoop.hbase.thrift.ThriftServer +conf, doAsEnabled, hbaseServiceHandler, host, httpEnabled, httpServer, infoServer, listenPort, metrics, pau seMonitor, processor, qop, securityEnabled, serviceUGI, tserver + @@ -190,7 +172,7 @@ implements org.apache.hadoop.util.Tool Constructor and Description -ThriftServer() +ThriftServer(org.apache.hadoop.conf.Configurationconf) @@ -208,140 +190,45 @@ implements org.apache.hadoop.util.Tool Method and Description -private static https://docs.oracle.com/javase/8/docs/api/java/net/InetSocketAddress.html?is-external=true; title="class or interface in java.net">InetSocketAddress -bindToPort(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringbindValue, - intlistenPort) +protected void +addOptions(org.apache.hbase.thirdparty.org.apache.commons.cli.Optionsoptions) +Add options to command lines + -private static https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorService -createExecutor(intworkerThreads, - intmaxCallQueueSize, - ThriftMetricsmetrics) +protected HBaseServiceHandler +createHandler(org.apache.hadoop.conf.Configurationconf, + UserProvideruserProvider) -private https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String -getBindAddress(org.apache.hadoop.conf.Configurationconf, - org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmd) +protected org.apache.thrift.TProcessor +createProcessor() -private https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String -getImplType(booleannonblocking, - booleanhsha, - booleanselector) - - -private int -getListenPort(org.apache.hadoop.conf.Configurationconf, - org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmd) - - -private static org.apache.hbase.thirdparty.org.apache.commons.cli.Options -getOptions() - - -private int
[31/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/CompleteScanResultCache.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/CompleteScanResultCache.html b/devapidocs/org/apache/hadoop/hbase/client/CompleteScanResultCache.html index 3442769..81f51d1 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/CompleteScanResultCache.html +++ b/devapidocs/org/apache/hadoop/hbase/client/CompleteScanResultCache.html @@ -442,6 +442,6 @@ implements Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/Connection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/Connection.html b/devapidocs/org/apache/hadoop/hbase/client/Connection.html index 8df693a..7edc55f 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/Connection.html +++ b/devapidocs/org/apache/hadoop/hbase/client/Connection.html @@ -530,6 +530,6 @@ extends Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/ConnectionConfiguration.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/ConnectionConfiguration.html b/devapidocs/org/apache/hadoop/hbase/client/ConnectionConfiguration.html index acd4ab0..cb2cf06 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/ConnectionConfiguration.html +++ b/devapidocs/org/apache/hadoop/hbase/client/ConnectionConfiguration.html @@ -864,6 +864,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/ConnectionFactory.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/ConnectionFactory.html b/devapidocs/org/apache/hadoop/hbase/client/ConnectionFactory.html index c725c56..4974699 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/ConnectionFactory.html +++ b/devapidocs/org/apache/hadoop/hbase/client/ConnectionFactory.html @@ -640,6 +640,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html b/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html index 990213e..6e32149 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html +++ b/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html @@ -382,6 +382,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html b/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html index 2827ff7..2833869 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html +++ b/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html @@ -329,6 +329,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved.
[31/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html index 6424016..40f83ba 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -static class HRegion.WriteState +static class HRegion.WriteState extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -239,7 +239,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html flushing -volatileboolean flushing +volatileboolean flushing @@ -248,7 +248,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html flushRequested -volatileboolean flushRequested +volatileboolean flushRequested @@ -257,7 +257,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html compacting -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicInteger compacting +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicInteger compacting @@ -266,7 +266,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html writesEnabled -volatileboolean writesEnabled +volatileboolean writesEnabled @@ -275,7 +275,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html readOnly -volatileboolean readOnly +volatileboolean readOnly @@ -284,7 +284,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html readsEnabled -volatileboolean readsEnabled +volatileboolean readsEnabled @@ -293,7 +293,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html HEAP_SIZE -static finallong HEAP_SIZE +static finallong HEAP_SIZE @@ -310,7 +310,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html WriteState -WriteState() +WriteState() @@ -327,7 +327,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html setReadOnly -voidsetReadOnly(booleanonOff) +voidsetReadOnly(booleanonOff) Set flags that make this region read-only. Parameters: @@ -341,7 +341,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html isReadOnly -booleanisReadOnly() +booleanisReadOnly() @@ -350,7 +350,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html isFlushRequested -booleanisFlushRequested() +booleanisFlushRequested() @@ -359,7 +359,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html setReadsEnabled -voidsetReadsEnabled(booleanreadsEnabled) +voidsetReadsEnabled(booleanreadsEnabled)
[31/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html index b062755..1ed3578 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; -static class HRegion.MutationBatchOperation +static class HRegion.MutationBatchOperation extends HRegion.BatchOperationMutation Batch of mutation operations. Base class is shared with HRegion.ReplayBatchOperation as most of the logic is same. @@ -342,7 +342,7 @@ extends nonceGroup -privatelong nonceGroup +privatelong nonceGroup @@ -351,7 +351,7 @@ extends nonce -privatelong nonce +privatelong nonce @@ -368,7 +368,7 @@ extends MutationBatchOperation -publicMutationBatchOperation(HRegionregion, +publicMutationBatchOperation(HRegionregion, Mutation[]operations, booleanatomic, longnonceGroup, @@ -389,7 +389,7 @@ extends getMutation -publicMutationgetMutation(intindex) +publicMutationgetMutation(intindex) Specified by: getMutationin classHRegion.BatchOperationMutation @@ -402,7 +402,7 @@ extends getNonceGroup -publiclonggetNonceGroup(intindex) +publiclonggetNonceGroup(intindex) Specified by: getNonceGroupin classHRegion.BatchOperationMutation @@ -415,7 +415,7 @@ extends getNonce -publiclonggetNonce(intindex) +publiclonggetNonce(intindex) Specified by: getNoncein classHRegion.BatchOperationMutation @@ -428,7 +428,7 @@ extends getMutationsForCoprocs -publicMutation[]getMutationsForCoprocs() +publicMutation[]getMutationsForCoprocs() Description copied from class:HRegion.BatchOperation This method is potentially expensive and useful mostly for non-replay CP path. @@ -443,7 +443,7 @@ extends isInReplay -publicbooleanisInReplay() +publicbooleanisInReplay() Specified by: isInReplayin classHRegion.BatchOperationMutation @@ -456,7 +456,7 @@ extends getOrigLogSeqNum -publiclonggetOrigLogSeqNum() +publiclonggetOrigLogSeqNum() Specified by: getOrigLogSeqNumin classHRegion.BatchOperationMutation @@ -469,7 +469,7 @@ extends startRegionOperation -publicvoidstartRegionOperation() +publicvoidstartRegionOperation() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Specified by: @@ -485,7 +485,7 @@ extends closeRegionOperation -publicvoidcloseRegionOperation() +publicvoidcloseRegionOperation() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Specified by: @@ -501,7 +501,7 @@ extends checkAndPreparePut -publicvoidcheckAndPreparePut(Putp) +publicvoidcheckAndPreparePut(Putp) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from class:HRegion.BatchOperation Implement any Put request specific check and prepare logic here. Please refer to @@ -520,7 +520,7 @@ extends checkAndPrepare -publicvoidcheckAndPrepare() +publicvoidcheckAndPrepare() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from class:HRegion.BatchOperation Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs @@ -542,7 +542,7 @@ extends prepareMiniBatchOperations -publicvoidprepareMiniBatchOperations(MiniBatchOperationInProgressMutationminiBatchOp, +publicvoidprepareMiniBatchOperations(MiniBatchOperationInProgressMutationminiBatchOp, longtimestamp, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegion.RowLockacquiredRowLocks) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -563,7 +563,7 @@ extends buildWALEdits -publichttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairNonceKey,WALEditbuildWALEdits(MiniBatchOperationInProgressMutation ;miniBatchOp)
[31/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html index 2cdee19..e6bc675 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html @@ -110,2406 +110,2407 @@ 102import org.apache.hadoop.hbase.util.ForeignExceptionUtil; 103import org.apache.hadoop.hbase.util.Pair; 104import org.apache.hadoop.hbase.wal.AbstractFSWALProvider; -105import org.apache.yetus.audience.InterfaceAudience; -106import org.apache.zookeeper.KeeperException; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109 -110import org.apache.hbase.thirdparty.com.google.protobuf.RpcController; -111import org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; -112import org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations; -113 -114import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -115import org.apache.hadoop.hbase.shaded.protobuf.ResponseConverter; -116import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -117import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse; -118import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest; -119import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -120import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockHeartbeatRequest; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockHeartbeatResponse; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockRequest; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockResponse; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockService; -133import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos; -134import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -135import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -136import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -137import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -138import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -139import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -140import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -141import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -142import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest; -143import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse; -144import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -145import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -146import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest; -147import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse; -148import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest; -149import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse; -150import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -151import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse; -152import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest; -153import
[31/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/TableNamespaceManager.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/TableNamespaceManager.html b/devapidocs/org/apache/hadoop/hbase/master/TableNamespaceManager.html index d45e83e..f5a5289 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/TableNamespaceManager.html +++ b/devapidocs/org/apache/hadoop/hbase/master/TableNamespaceManager.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":9,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":9,"i5":9,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -107,22 +107,13 @@ var activeTableTab = "activeTableTab"; - -All Implemented Interfaces: -Stoppable - @InterfaceAudience.Private -public class TableNamespaceManager -extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object -implements Stoppable -This is a helper class used internally to manage the namespace metadata that is stored in - TableName.NAMESPACE_TABLE_NAME. It also mirrors updates to the ZK store by forwarding updates to - ZKNamespaceManager. - - WARNING: Do not use. Go via the higher-level ClusterSchema API instead. This manager - is likely to go aways anyways. +public class TableNamespaceManager +extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +This is a helper class used internally to manage the namespace metadata that is stored in the ns + family in meta table. @@ -142,29 +133,21 @@ implements Field and Description -private org.apache.hadoop.conf.Configuration -conf +private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,NamespaceDescriptor +cache (package private) static int DEFAULT_NS_INIT_TIMEOUT -private boolean -initialized - - static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String KEY_MAX_REGIONS - + static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String KEY_MAX_TABLES - -private static org.slf4j.Logger -LOG - private MasterServices masterServices @@ -173,18 +156,6 @@ implements (package private) static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String NS_INIT_TIMEOUT - -private Table -nsTable - - -private boolean -stopped - - -private ZKNamespaceManager -zkNamespaceManager - @@ -218,105 +189,53 @@ implements Method and Description -private static void -block(MasterServicesservices, - longprocId) -An ugly utility to be removed when refactor TableNamespaceManager. - +void +addOrUpdateNamespace(NamespaceDescriptorns) -private void -blockingCreateNamespace(NamespaceDescriptornamespaceDescriptor) -Create Namespace in a blocking manner. - +void +deleteNamespace(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringnamespaceName) -private void -createNamespaceTable(MasterServicesmasterServices) - - boolean -doesNamespaceExist(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringnamespaceName) +doesNamespaceExist(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringnamespaceName) +check whether a namespace has already existed. + - + NamespaceDescriptor get(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) - -private NamespaceDescriptor -get(Tabletable, - https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) - - + static long getMaxRegions(NamespaceDescriptorns) - + static long getMaxTables(NamespaceDescriptorns) - -private Table -getNamespaceTable() - - -private TableState -getTableState() - - -void -insertIntoNSTable(NamespaceDescriptorns) - - -boolean -isStopped() - - -private boolean
[31/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html b/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html index 9ca96d9..0a6df8c 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html +++ b/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html @@ -129,10 +129,6 @@ private Permission.Action AuthResult.action - -protected Permission.Action[] -Permission.actions - @@ -146,6 +142,10 @@ protected static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/Byte.html?is-external=true; title="class or interface in java.lang">Byte,Permission.Action Permission.ACTION_BY_CODE + +protected https://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true; title="class or interface in java.util">EnumSetPermission.Action +Permission.actions + @@ -220,6 +220,15 @@ the order they are declared. +boolean +AuthManager.accessUserTable(Useruser, + TableNametable, + Permission.Actionaction) +Checks if the user has access to the full table or at least a family/qualifier + for the specified action. + + + static AuthResult AuthResult.allow(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringrequest, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringreason, @@ -227,7 +236,7 @@ the order they are declared. Permission.Actionaction, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringnamespace) - + static AuthResult AuthResult.allow(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringrequest, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringreason, @@ -237,7 +246,7 @@ the order they are declared. byte[]family, byte[]qualifier) - + static AuthResult AuthResult.allow(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringrequest, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringreason, @@ -246,100 +255,90 @@ the order they are declared. TableNametable, https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],? extends https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true; title="class or interface in java.util">Collection?families) - -private boolean -TableAuthManager.authorize(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPermissionperms, - Permission.Actionaction) -Authorizes a global permission - - -private boolean -TableAuthManager.authorize(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTablePermissionperms, - https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringnamespace, - Permission.Actionaction) +boolean +AuthManager.authorizeCell(Useruser, + TableNametable, + Cellcell, + Permission.Actionaction) +Check if user has given action privilige in cell scope. + private boolean -TableAuthManager.authorize(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTablePermissionperms, - TableNametable, - byte[]family, - byte[]qualifier, - Permission.Actionaction) +AuthManager.authorizeFamily(https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetTablePermissionpermissions, + TableNametable, + byte[]family, + Permission.Actionaction) -boolean -TableAuthManager.authorize(Useruser, - Permission.Actionaction) -Authorize a global permission based on ACLs for the given user and the - user's groups. - +private boolean +AuthManager.authorizeGlobal(GlobalPermissionpermissions, + Permission.Actionaction) -boolean -TableAuthManager.authorize(Useruser, -
[31/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureRetainer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureRetainer.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureRetainer.html index 0af8acd..c5f21ac 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureRetainer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureRetainer.html @@ -645,1615 +645,1597 @@ 637 proc.afterReplay(getEnvironment()); 638 } 639}); -640 -641// 4. Push the procedures to the timeout executor -642waitingTimeoutList.forEach(proc - { -643 proc.afterReplay(getEnvironment()); -644 timeoutExecutor.add(proc); -645}); -646// 5. restore locks -647restoreLocks(); -648// 6. Push the procedure to the scheduler -649 failedList.forEach(scheduler::addBack); -650runnableList.forEach(p - { -651 p.afterReplay(getEnvironment()); -652 if (!p.hasParent()) { -653 sendProcedureLoadedNotification(p.getProcId()); -654 } -655 // If the procedure holds the lock, put the procedure in front -656 // If its parent holds the lock, put the procedure in front -657 // TODO. Is that possible that its ancestor holds the lock? -658 // For now, the deepest procedure hierarchy is: -659 // ModifyTableProcedure - ReopenTableProcedure - -660 // MoveTableProcedure - Unassign/AssignProcedure -661 // But ModifyTableProcedure and ReopenTableProcedure won't hold the lock -662 // So, check parent lock is enough(a tricky case is resovled by HBASE-21384). -663 // If some one change or add new procedures making 'grandpa' procedure -664 // holds the lock, but parent procedure don't hold the lock, there will -665 // be a problem here. We have to check one procedure's ancestors. -666 // And we need to change LockAndQueue.hasParentLock(Procedure? proc) method -667 // to check all ancestors too. -668 if (p.isLockedWhenLoading() || (p.hasParent() procedures -669 .get(p.getParentProcId()).isLockedWhenLoading())) { -670scheduler.addFront(p, false); -671 } else { -672// if it was not, it can wait. -673scheduler.addBack(p, false); -674 } -675}); -676// After all procedures put into the queue, signal the worker threads. -677// Otherwise, there is a race condition. See HBASE-21364. -678scheduler.signalAll(); -679 } +640// 4. restore locks +641restoreLocks(); +642 +643// 5. Push the procedures to the timeout executor +644waitingTimeoutList.forEach(proc - { +645 proc.afterReplay(getEnvironment()); +646 timeoutExecutor.add(proc); +647}); +648 +649// 6. Push the procedure to the scheduler +650 failedList.forEach(scheduler::addBack); +651runnableList.forEach(p - { +652 p.afterReplay(getEnvironment()); +653 if (!p.hasParent()) { +654 sendProcedureLoadedNotification(p.getProcId()); +655 } +656 scheduler.addBack(p); +657}); +658// After all procedures put into the queue, signal the worker threads. +659// Otherwise, there is a race condition. See HBASE-21364. +660scheduler.signalAll(); +661 } +662 +663 /** +664 * Initialize the procedure executor, but do not start workers. We will start them later. +665 * p/ +666 * It calls ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, and +667 * ensure a single executor, and start the procedure replay to resume and recover the previous +668 * pending and in-progress procedures. +669 * @param numThreads number of threads available for procedure execution. +670 * @param abortOnCorruption true if you want to abort your service in case a corrupted procedure +671 * is found on replay. otherwise false. +672 */ +673 public void init(int numThreads, boolean abortOnCorruption) throws IOException { +674// We have numThreads executor + one timer thread used for timing out +675// procedures and triggering periodic procedures. +676this.corePoolSize = numThreads; +677this.maxPoolSize = 10 * numThreads; +678LOG.info("Starting {} core workers (bigger of cpus/4 or 16) with max (burst) worker count={}", +679corePoolSize, maxPoolSize); 680 -681 /** -682 * Initialize the procedure executor, but do not start workers. We will start them later. -683 * p/ -684 * It calls ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, and -685 * ensure a single executor, and start the procedure replay to resume and recover the previous -686 * pending and
[31/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LoadCandidateGenerator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LoadCandidateGenerator.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LoadCandidateGenerator.html index c7d99b2..9d1542c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LoadCandidateGenerator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LoadCandidateGenerator.html @@ -382,1357 +382,1365 @@ 374for (int i = 0; i this.curFunctionCosts.length; i++) { 375 curFunctionCosts[i] = tempFunctionCosts[i]; 376} -377LOG.info("start StochasticLoadBalancer.balancer, initCost=" + currentCost + ", functionCost=" -378+ functionCost()); +377double initCost = currentCost; +378double newCost = currentCost; 379 -380double initCost = currentCost; -381double newCost = currentCost; -382 -383long computedMaxSteps; -384if (runMaxSteps) { -385 computedMaxSteps = Math.max(this.maxSteps, -386 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -387} else { -388 computedMaxSteps = Math.min(this.maxSteps, -389 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -390} -391// Perform a stochastic walk to see if we can get a good fit. -392long step; -393 -394for (step = 0; step computedMaxSteps; step++) { -395 Cluster.Action action = nextAction(cluster); -396 -397 if (action.type == Type.NULL) { -398continue; -399 } -400 -401 cluster.doAction(action); -402 updateCostsWithAction(cluster, action); -403 -404 newCost = computeCost(cluster, currentCost); -405 -406 // Should this be kept? -407 if (newCost currentCost) { -408currentCost = newCost; -409 -410// save for JMX -411curOverallCost = currentCost; -412for (int i = 0; i this.curFunctionCosts.length; i++) { -413 curFunctionCosts[i] = tempFunctionCosts[i]; -414} -415 } else { -416// Put things back the way they were before. -417// TODO: undo by remembering old values -418Action undoAction = action.undoAction(); -419cluster.doAction(undoAction); -420updateCostsWithAction(cluster, undoAction); -421 } -422 -423 if (EnvironmentEdgeManager.currentTime() - startTime -424 maxRunningTime) { -425break; -426 } -427} -428long endTime = EnvironmentEdgeManager.currentTime(); -429 -430 metricsBalancer.balanceCluster(endTime - startTime); -431 -432// update costs metrics -433updateStochasticCosts(tableName, curOverallCost, curFunctionCosts); -434if (initCost currentCost) { -435 plans = createRegionPlans(cluster); -436 LOG.info("Finished computing new load balance plan. Computation took {}" + -437" to try {} different iterations. Found a solution that moves " + -438"{} regions; Going from a computed cost of {}" + -439" to a new cost of {}", java.time.Duration.ofMillis(endTime - startTime), -440step, plans.size(), initCost, currentCost); -441 return plans; -442} -443LOG.info("Could not find a better load balance plan. Tried {} different configurations in " + -444 "{}, and did not find anything with a computed cost less than {}", step, -445 java.time.Duration.ofMillis(endTime - startTime), initCost); -446return null; -447 } -448 -449 /** -450 * update costs to JMX -451 */ -452 private void updateStochasticCosts(TableName tableName, Double overall, Double[] subCosts) { -453if (tableName == null) return; -454 -455// check if the metricsBalancer is MetricsStochasticBalancer before casting -456if (metricsBalancer instanceof MetricsStochasticBalancer) { -457 MetricsStochasticBalancer balancer = (MetricsStochasticBalancer) metricsBalancer; -458 // overall cost -459 balancer.updateStochasticCost(tableName.getNameAsString(), -460"Overall", "Overall cost", overall); -461 -462 // each cost function -463 for (int i = 0; i costFunctions.length; i++) { -464CostFunction costFunction = costFunctions[i]; -465String costFunctionName = costFunction.getClass().getSimpleName(); -466Double costPercent = (overall == 0) ? 0 : (subCosts[i] / overall); -467// TODO: cost function may need a specific description -468 balancer.updateStochasticCost(tableName.getNameAsString(), costFunctionName, -469 "The percent of " + costFunctionName, costPercent); -470 } -471} -472 } -473
[31/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.html index e4dc134..2945c58 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.html @@ -93,157 +93,163 @@ 085boolean hasNext(); 086 087/** -088 * @return true if the iterator next element is a completed procedure. -089 */ -090boolean isNextFinished(); -091 -092/** -093 * Skip the next procedure -094 */ -095void skipNext(); -096 -097/** -098 * Returns the next procedure in the iteration. -099 * @throws IOException if there was an error fetching/deserializing the procedure -100 * @return the next procedure in the iteration. -101 */ -102@SuppressWarnings("rawtypes") -103Procedure next() throws IOException; -104 } -105 -106 /** -107 * Interface passed to the ProcedureStore.load() method to handle the store-load events. -108 */ -109 public interface ProcedureLoader { -110/** -111 * Called by ProcedureStore.load() to notify about the maximum proc-id in the store. -112 * @param maxProcId the highest proc-id in the store -113 */ -114void setMaxProcId(long maxProcId); -115 +088 * Calling this method does not need to converting the protobuf message to the Procedure class, +089 * so if it returns true we can call {@link #skipNext()} to skip the procedure without +090 * deserializing. This could increase the performance. +091 * @return true if the iterator next element is a completed procedure. +092 */ +093boolean isNextFinished(); +094 +095/** +096 * Skip the next procedure +097 * p/ +098 * This method is used to skip the deserializing of the procedure to increase performance, as +099 * when calling next we need to convert the protobuf message to the Procedure class. +100 */ +101void skipNext(); +102 +103/** +104 * Returns the next procedure in the iteration. +105 * @throws IOException if there was an error fetching/deserializing the procedure +106 * @return the next procedure in the iteration. +107 */ +108@SuppressWarnings("rawtypes") +109Procedure next() throws IOException; +110 } +111 +112 /** +113 * Interface passed to the ProcedureStore.load() method to handle the store-load events. +114 */ +115 public interface ProcedureLoader { 116/** -117 * Called by the ProcedureStore.load() every time a set of procedures are ready to be executed. -118 * The ProcedureIterator passed to the method, has the procedure sorted in replay-order. -119 * @param procIter iterator over the procedures ready to be added to the executor. -120 */ -121void load(ProcedureIterator procIter) throws IOException; -122 -123/** -124 * Called by the ProcedureStore.load() in case we have procedures not-ready to be added to -125 * the executor, which probably means they are corrupted since some information/link is missing. -126 * @param procIter iterator over the procedures not ready to be added to the executor, corrupted -127 */ -128void handleCorrupted(ProcedureIterator procIter) throws IOException; -129 } -130 -131 /** -132 * Add the listener to the notification list. -133 * @param listener The AssignmentListener to register -134 */ -135 void registerListener(ProcedureStoreListener listener); +117 * Called by ProcedureStore.load() to notify about the maximum proc-id in the store. +118 * @param maxProcId the highest proc-id in the store +119 */ +120void setMaxProcId(long maxProcId); +121 +122/** +123 * Called by the ProcedureStore.load() every time a set of procedures are ready to be executed. +124 * The ProcedureIterator passed to the method, has the procedure sorted in replay-order. +125 * @param procIter iterator over the procedures ready to be added to the executor. +126 */ +127void load(ProcedureIterator procIter) throws IOException; +128 +129/** +130 * Called by the ProcedureStore.load() in case we have procedures not-ready to be added to +131 * the executor, which probably means they are corrupted since some information/link is missing. +132 * @param procIter iterator over the procedures not ready to be added to the executor, corrupted +133 */ +134void handleCorrupted(ProcedureIterator procIter) throws IOException; +135 } 136 137 /** -138 * Remove the listener from the notification list. -139 * @param listener The AssignmentListener to unregister -140 * @return true if the listner
[31/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html index 7c9bc39..f9d1cea 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html @@ -351,7 +351,7 @@ extends Procedure -acquireLock, addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doExecute, doRollba ck, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcId, g etProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, releaseLock, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setSt ackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, skipPersistence, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted +acquireLock, addStackIndex, afterReplay, beforeReplay, bypass, compareTo, completionCleanup, doExecute, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProc Id, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, releaseLock, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, set Failure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, skipPersistence, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStateNode.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStateNode.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStateNode.html index f8fa74a..a3e6f22 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStateNode.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStateNode.html @@ -265,8 +265,9 @@ private TransitRegionStateProcedure -AssignmentManager.createAssignProcedure(RegionStateNoderegionNode, - ServerNametargetServer) +AssignmentManager.createAssignProcedure(RegionStateNoderegionNode, + ServerNametargetServer, + booleanoverride) private HRegionLocation @@ -278,82 +279,87 @@ https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforesult) +private TransitRegionStateProcedure +AssignmentManager.createUnassignProcedure(RegionStateNoderegionNode, + booleanoverride) + + protected void UnassignProcedure.finishTransition(MasterProcedureEnvenv, RegionStateNoderegionNode) Deprecated. - + protected void AssignProcedure.finishTransition(MasterProcedureEnvenv, RegionStateNoderegionNode) Deprecated. - + protected abstract void RegionTransitionProcedure.finishTransition(MasterProcedureEnvenv, RegionStateNoderegionNode) Deprecated. - + (package private) boolean RegionStates.include(RegionStateNodenode, booleanoffline) Utility. - + private boolean TransitRegionStateProcedure.incrementAndCheckMaxAttempts(MasterProcedureEnvenv,
[31/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreTracker.DeleteState.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreTracker.DeleteState.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreTracker.DeleteState.html index 0c69df9..8515fa3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreTracker.DeleteState.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreTracker.DeleteState.html @@ -32,263 +32,263 @@ 024import java.util.Map; 025import java.util.TreeMap; 026import java.util.stream.LongStream; -027import org.apache.yetus.audience.InterfaceAudience; -028import org.apache.yetus.audience.InterfaceStability; -029 -030import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos; -031 -032/** -033 * Keeps track of live procedures. -034 * -035 * It can be used by the ProcedureStore to identify which procedures are already -036 * deleted/completed to avoid the deserialization step on restart -037 */ -038@InterfaceAudience.Private -039@InterfaceStability.Evolving -040public class ProcedureStoreTracker { -041 // Key is procedure id corresponding to first bit of the bitmap. -042 private final TreeMapLong, BitSetNode map = new TreeMap(); -043 -044 /** -045 * If true, do not remove bits corresponding to deleted procedures. Note that this can result -046 * in huge bitmaps overtime. -047 * Currently, it's set to true only when building tracker state from logs during recovery. During -048 * recovery, if we are sure that a procedure has been deleted, reading its old update entries -049 * can be skipped. -050 */ -051 private boolean keepDeletes = false; -052 /** -053 * If true, it means tracker has incomplete information about the active/deleted procedures. -054 * It's set to true only when recovering from old logs. See {@link #isDeleted(long)} docs to -055 * understand it's real use. -056 */ -057 boolean partial = false; -058 -059 private long minModifiedProcId = Long.MAX_VALUE; -060 private long maxModifiedProcId = Long.MIN_VALUE; -061 -062 public enum DeleteState { YES, NO, MAYBE } -063 -064 public void resetToProto(ProcedureProtos.ProcedureStoreTracker trackerProtoBuf) { -065reset(); -066for (ProcedureProtos.ProcedureStoreTracker.TrackerNode protoNode: trackerProtoBuf.getNodeList()) { -067 final BitSetNode node = new BitSetNode(protoNode); -068 map.put(node.getStart(), node); -069} -070 } -071 -072 /** -073 * Resets internal state to same as given {@code tracker}. Does deep copy of the bitmap. -074 */ -075 public void resetTo(ProcedureStoreTracker tracker) { -076resetTo(tracker, false); -077 } -078 -079 /** -080 * Resets internal state to same as given {@code tracker}, and change the deleted flag according -081 * to the modified flag if {@code resetDelete} is true. Does deep copy of the bitmap. -082 * p/ -083 * The {@code resetDelete} will be set to true when building cleanup tracker, please see the -084 * comments in {@link BitSetNode#BitSetNode(BitSetNode, boolean)} to learn how we change the -085 * deleted flag if {@code resetDelete} is true. -086 */ -087 public void resetTo(ProcedureStoreTracker tracker, boolean resetDelete) { -088reset(); -089this.partial = tracker.partial; -090this.minModifiedProcId = tracker.minModifiedProcId; -091this.maxModifiedProcId = tracker.maxModifiedProcId; -092this.keepDeletes = tracker.keepDeletes; -093for (Map.EntryLong, BitSetNode entry : tracker.map.entrySet()) { -094 map.put(entry.getKey(), new BitSetNode(entry.getValue(), resetDelete)); -095} -096 } -097 -098 public void insert(long procId) { -099insert(null, procId); -100 } -101 -102 public void insert(long[] procIds) { -103for (int i = 0; i procIds.length; ++i) { -104 insert(procIds[i]); -105} -106 } -107 -108 public void insert(long procId, long[] subProcIds) { -109BitSetNode node = update(null, procId); -110for (int i = 0; i subProcIds.length; ++i) { -111 node = insert(node, subProcIds[i]); -112} -113 } -114 -115 private BitSetNode insert(BitSetNode node, long procId) { -116if (node == null || !node.contains(procId)) { -117 node = getOrCreateNode(procId); -118} -119node.insertOrUpdate(procId); -120trackProcIds(procId); -121return node; -122 } -123 -124 public void update(long procId) { -125update(null, procId); -126 } -127 -128 private BitSetNode update(BitSetNode node, long procId) { -129node = lookupClosestNode(node, procId); -130assert node != null : "expected node to update procId=" + procId; -131assert node.contains(procId) : "expected procId=" + procId + " in
[31/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html index 061ce80..bdfc3f8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html @@ -39,2126 +39,2163 @@ 031import java.util.Set; 032import java.util.concurrent.ConcurrentHashMap; 033import java.util.concurrent.CopyOnWriteArrayList; -034import java.util.concurrent.TimeUnit; -035import java.util.concurrent.atomic.AtomicBoolean; -036import java.util.concurrent.atomic.AtomicInteger; -037import java.util.concurrent.atomic.AtomicLong; -038import java.util.stream.Collectors; -039import java.util.stream.Stream; -040import org.apache.hadoop.conf.Configuration; -041import org.apache.hadoop.hbase.HConstants; -042import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; -043import org.apache.hadoop.hbase.log.HBaseMarkers; -044import org.apache.hadoop.hbase.procedure2.Procedure.LockState; -045import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; -046import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; -047import org.apache.hadoop.hbase.procedure2.util.StringUtils; -048import org.apache.hadoop.hbase.security.User; -049import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -050import org.apache.hadoop.hbase.util.IdLock; -051import org.apache.hadoop.hbase.util.NonceKey; -052import org.apache.hadoop.hbase.util.Threads; -053import org.apache.yetus.audience.InterfaceAudience; -054import org.slf4j.Logger; -055import org.slf4j.LoggerFactory; -056 -057import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -058import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +034import java.util.concurrent.Executor; +035import java.util.concurrent.Executors; +036import java.util.concurrent.TimeUnit; +037import java.util.concurrent.atomic.AtomicBoolean; +038import java.util.concurrent.atomic.AtomicInteger; +039import java.util.concurrent.atomic.AtomicLong; +040import java.util.stream.Collectors; +041import java.util.stream.Stream; +042import org.apache.hadoop.conf.Configuration; +043import org.apache.hadoop.hbase.HConstants; +044import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; +045import org.apache.hadoop.hbase.log.HBaseMarkers; +046import org.apache.hadoop.hbase.procedure2.Procedure.LockState; +047import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; +048import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; +049import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureStoreListener; +050import org.apache.hadoop.hbase.procedure2.util.StringUtils; +051import org.apache.hadoop.hbase.security.User; +052import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +053import org.apache.hadoop.hbase.util.IdLock; +054import org.apache.hadoop.hbase.util.NonceKey; +055import org.apache.hadoop.hbase.util.Threads; +056import org.apache.yetus.audience.InterfaceAudience; +057import org.slf4j.Logger; +058import org.slf4j.LoggerFactory; 059 -060import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; -061 -062/** -063 * Thread Pool that executes the submitted procedures. -064 * The executor has a ProcedureStore associated. -065 * Each operation is logged and on restart the pending procedures are resumed. -066 * -067 * Unless the Procedure code throws an error (e.g. invalid user input) -068 * the procedure will complete (at some point in time), On restart the pending -069 * procedures are resumed and the once failed will be rolledback. +060import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +061import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +062import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; +063 +064import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; +065 +066/** +067 * Thread Pool that executes the submitted procedures. +068 * The executor has a ProcedureStore associated. +069 * Each operation is logged and on restart the pending procedures are resumed. 070 * -071 * The user can add procedures to the executor via submitProcedure(proc) -072 * check for the finished state via isFinished(procId) -073 * and get the result via getResult(procId) -074 */ -075@InterfaceAudience.Private -076public class ProcedureExecutorTEnvironment {
[31/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html b/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html index b05e4c3..773eb00 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/TwoPhaseProcedure.html @@ -180,7 +180,7 @@ extends Procedure -abort, acquireLock, addStackIndex, afterReplay, beforeReplay, bypass, compareTo, completionCleanup, deserializeStateData, doAcquireLock, doExecute, doReleaseLock, doRollback, elapsedTime, execute, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutT imestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, lockedWhenLoading, releaseLock, removeStackIndex, restoreLock, rollback, serializeStateData, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcI d, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, tryRunnable, updateMetricsOnF inish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted +abort, acquireLock, addStackIndex, afterReplay, beforeReplay, bypass, compareTo, completionCleanup, deserializeStateData, doAcquireLock, doExecute, doReleaseLock, doRollback, elapsedTime, execute, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutT imestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, lockedWhenLoading, needPersistence, releaseLock, removeStackIndex, resetPersistence, restoreLock, rollback, serializeStateData, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, s etOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, skipPersistence, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, tryRunnable, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureScheduler.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureScheduler.html b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureScheduler.html index 795dd52..9a15669 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureScheduler.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureScheduler.html @@ -162,6 +162,10 @@ (package private) ProcedureScheduler +ProcedureExecutor.getProcedureScheduler() + + +(package private) ProcedureScheduler ProcedureExecutor.getScheduler() http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html index 525817b..fd4a0a8 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureSuspendedException.html @@ -236,6 +236,19 @@ Uses of ProcedureSuspendedException in org.apache.hadoop.hbase.master.replication +Methods in org.apache.hadoop.hbase.master.replication that return ProcedureSuspendedException + +Modifier and Type +Method and Description + + + +private
[31/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html index 0d425bd..4075394 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.WriteState.html @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -static class HRegion.WriteState +static class HRegion.WriteState extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -239,7 +239,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html flushing -volatileboolean flushing +volatileboolean flushing @@ -248,7 +248,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html flushRequested -volatileboolean flushRequested +volatileboolean flushRequested @@ -257,7 +257,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html compacting -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicInteger compacting +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicInteger compacting @@ -266,7 +266,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html writesEnabled -volatileboolean writesEnabled +volatileboolean writesEnabled @@ -275,7 +275,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html readOnly -volatileboolean readOnly +volatileboolean readOnly @@ -284,7 +284,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html readsEnabled -volatileboolean readsEnabled +volatileboolean readsEnabled @@ -293,7 +293,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html HEAP_SIZE -static finallong HEAP_SIZE +static finallong HEAP_SIZE @@ -310,7 +310,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html WriteState -WriteState() +WriteState() @@ -327,7 +327,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html setReadOnly -voidsetReadOnly(booleanonOff) +voidsetReadOnly(booleanonOff) Set flags that make this region read-only. Parameters: @@ -341,7 +341,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html isReadOnly -booleanisReadOnly() +booleanisReadOnly() @@ -350,7 +350,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html isFlushRequested -booleanisFlushRequested() +booleanisFlushRequested() @@ -359,7 +359,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html setReadsEnabled -voidsetReadsEnabled(booleanreadsEnabled) +voidsetReadsEnabled(booleanreadsEnabled)
[31/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.PureJavaComparer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.PureJavaComparer.html b/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.PureJavaComparer.html index 3d721e1..d4d6c23 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.PureJavaComparer.html +++ b/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.PureJavaComparer.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -static enum Bytes.LexicographicalComparerHolder.PureJavaComparer +static enum Bytes.LexicographicalComparerHolder.PureJavaComparer extends https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumBytes.LexicographicalComparerHolder.PureJavaComparer implements Bytes.Comparerbyte[] @@ -217,7 +217,7 @@ the order they are declared. INSTANCE -public static finalBytes.LexicographicalComparerHolder.PureJavaComparer INSTANCE +public static finalBytes.LexicographicalComparerHolder.PureJavaComparer INSTANCE @@ -234,7 +234,7 @@ the order they are declared. values -public staticBytes.LexicographicalComparerHolder.PureJavaComparer[]values() +public staticBytes.LexicographicalComparerHolder.PureJavaComparer[]values() Returns an array containing the constants of this enum type, in the order they are declared. This method may be used to iterate over the constants as follows: @@ -254,7 +254,7 @@ for (Bytes.LexicographicalComparerHolder.PureJavaComparer c : Bytes.Lexicographi valueOf -public staticBytes.LexicographicalComparerHolder.PureJavaComparervalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) +public staticBytes.LexicographicalComparerHolder.PureJavaComparervalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Returns the enum constant of this type with the specified name. The string must match exactly an identifier used to declare an enum constant in this type. (Extraneous whitespace characters are @@ -276,7 +276,7 @@ not permitted.) compareTo -publicintcompareTo(byte[]buffer1, +publicintcompareTo(byte[]buffer1, intoffset1, intlength1, byte[]buffer2, http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.UnsafeComparer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.UnsafeComparer.html b/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.UnsafeComparer.html index 909bc09..5b19db9 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.UnsafeComparer.html +++ b/devapidocs/org/apache/hadoop/hbase/util/Bytes.LexicographicalComparerHolder.UnsafeComparer.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -static enum Bytes.LexicographicalComparerHolder.UnsafeComparer +static enum Bytes.LexicographicalComparerHolder.UnsafeComparer extends https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumBytes.LexicographicalComparerHolder.UnsafeComparer implements Bytes.Comparerbyte[] @@ -238,7 +238,7 @@ the order they are declared. INSTANCE -public static finalBytes.LexicographicalComparerHolder.UnsafeComparer INSTANCE +public static finalBytes.LexicographicalComparerHolder.UnsafeComparer INSTANCE @@ -255,7 +255,7 @@ the order they are declared. theUnsafe -static finalsun.misc.Unsafe theUnsafe +static finalsun.misc.Unsafe theUnsafe @@ -272,7 +272,7 @@ the order they are declared. values -public staticBytes.LexicographicalComparerHolder.UnsafeComparer[]values() +public staticBytes.LexicographicalComparerHolder.UnsafeComparer[]values() Returns an array containing the constants of this enum type, in the order they are declared. This method may be used to iterate over the constants as follows: @@ -292,7 +292,7 @@ for (Bytes.LexicographicalComparerHolder.UnsafeComparer c : Bytes.Lexicographica valueOf -public staticBytes.LexicographicalComparerHolder.UnsafeComparervalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) +public staticBytes.LexicographicalComparerHolder.UnsafeComparervalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
[31/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyColumnFamilyFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyColumnFamilyFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyColumnFamilyFuture.html index a5789e0..93a57cb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyColumnFamilyFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyColumnFamilyFuture.html @@ -238,4120 +238,4119 @@ 230 * @see Admin 231 */ 232@InterfaceAudience.Private -233@InterfaceStability.Evolving -234public class HBaseAdmin implements Admin { -235 private static final Logger LOG = LoggerFactory.getLogger(HBaseAdmin.class); -236 -237 private ClusterConnection connection; -238 -239 private final Configuration conf; -240 private final long pause; -241 private final int numRetries; -242 private final int syncWaitTimeout; -243 private boolean aborted; -244 private int operationTimeout; -245 private int rpcTimeout; -246 -247 private RpcRetryingCallerFactory rpcCallerFactory; -248 private RpcControllerFactory rpcControllerFactory; -249 -250 private NonceGenerator ng; -251 -252 @Override -253 public int getOperationTimeout() { -254return operationTimeout; -255 } -256 -257 HBaseAdmin(ClusterConnection connection) throws IOException { -258this.conf = connection.getConfiguration(); -259this.connection = connection; -260 -261// TODO: receive ConnectionConfiguration here rather than re-parsing these configs every time. -262this.pause = this.conf.getLong(HConstants.HBASE_CLIENT_PAUSE, -263 HConstants.DEFAULT_HBASE_CLIENT_PAUSE); -264this.numRetries = this.conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, -265 HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER); -266this.operationTimeout = this.conf.getInt(HConstants.HBASE_CLIENT_OPERATION_TIMEOUT, -267 HConstants.DEFAULT_HBASE_CLIENT_OPERATION_TIMEOUT); -268this.rpcTimeout = this.conf.getInt(HConstants.HBASE_RPC_TIMEOUT_KEY, -269 HConstants.DEFAULT_HBASE_RPC_TIMEOUT); -270this.syncWaitTimeout = this.conf.getInt( -271 "hbase.client.sync.wait.timeout.msec", 10 * 6); // 10min -272 -273this.rpcCallerFactory = connection.getRpcRetryingCallerFactory(); -274this.rpcControllerFactory = connection.getRpcControllerFactory(); -275 -276this.ng = this.connection.getNonceGenerator(); -277 } -278 -279 @Override -280 public void abort(String why, Throwable e) { -281// Currently does nothing but throw the passed message and exception -282this.aborted = true; -283throw new RuntimeException(why, e); -284 } -285 -286 @Override -287 public boolean isAborted() { -288return this.aborted; -289 } -290 -291 @Override -292 public boolean abortProcedure(final long procId, final boolean mayInterruptIfRunning) -293 throws IOException { -294return get(abortProcedureAsync(procId, mayInterruptIfRunning), this.syncWaitTimeout, -295 TimeUnit.MILLISECONDS); -296 } -297 -298 @Override -299 public FutureBoolean abortProcedureAsync(final long procId, final boolean mayInterruptIfRunning) -300 throws IOException { -301Boolean abortProcResponse = -302executeCallable(new MasterCallableAbortProcedureResponse(getConnection(), -303getRpcControllerFactory()) { -304 @Override -305 protected AbortProcedureResponse rpcCall() throws Exception { -306AbortProcedureRequest abortProcRequest = -307 AbortProcedureRequest.newBuilder().setProcId(procId).build(); -308return master.abortProcedure(getRpcController(), abortProcRequest); -309 } -310}).getIsProcedureAborted(); -311return new AbortProcedureFuture(this, procId, abortProcResponse); -312 } -313 -314 @Override -315 public ListTableDescriptor listTableDescriptors() throws IOException { -316return listTableDescriptors((Pattern)null, false); -317 } -318 -319 @Override -320 public ListTableDescriptor listTableDescriptors(Pattern pattern) throws IOException { -321return listTableDescriptors(pattern, false); -322 } -323 -324 @Override -325 public ListTableDescriptor listTableDescriptors(Pattern pattern, boolean includeSysTables) -326 throws IOException { -327return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -328getRpcControllerFactory()) { -329 @Override -330 protected ListTableDescriptor rpcCall() throws Exception { -331GetTableDescriptorsRequest req = -332 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables); -333return ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(), -334req)); -335 }
[31/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html b/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html index 35007fe..f6fd78b 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html +++ b/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html @@ -120,7 +120,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -class VisibilityLabelFilter +class VisibilityLabelFilter extends FilterBase This Filter checks the visibility expression with each KV against visibility labels associated with the scan. Based on the check the KV is included in the scan result or gets filtered out. @@ -290,7 +290,7 @@ extends expEvaluator -private finalVisibilityExpEvaluator expEvaluator +private finalVisibilityExpEvaluator expEvaluator @@ -299,7 +299,7 @@ extends cfVsMaxVersions -private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapByteRange,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer cfVsMaxVersions +private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapByteRange,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer cfVsMaxVersions @@ -308,7 +308,7 @@ extends curFamily -private finalByteRange curFamily +private finalByteRange curFamily @@ -317,7 +317,7 @@ extends curQualifier -private finalByteRange curQualifier +private finalByteRange curQualifier @@ -326,7 +326,7 @@ extends curFamilyMaxVersions -privateint curFamilyMaxVersions +privateint curFamilyMaxVersions @@ -335,7 +335,7 @@ extends curQualMetVersions -privateint curQualMetVersions +privateint curQualMetVersions @@ -352,7 +352,7 @@ extends VisibilityLabelFilter -publicVisibilityLabelFilter(VisibilityExpEvaluatorexpEvaluator, +publicVisibilityLabelFilter(VisibilityExpEvaluatorexpEvaluator, https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapByteRange,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">IntegercfVsMaxVersions) @@ -370,7 +370,7 @@ extends filterRowKey -publicbooleanfilterRowKey(Cellcell) +publicbooleanfilterRowKey(Cellcell) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from class:Filter Filters a row based on the row key. If this returns true, the entire row will be excluded. If @@ -398,7 +398,7 @@ extends filterCell -publicFilter.ReturnCodefilterCell(Cellcell) +publicFilter.ReturnCodefilterCell(Cellcell) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from class:Filter A way to filter based on the column family, column qualifier and/or the column value. Return @@ -435,7 +435,7 @@ extends reset -publicvoidreset() +publicvoidreset() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from class:FilterBase Filters that are purely stateless and do nothing in their reset() methods can inherit http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html index 4ade4c1..03e1383 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html @@ -199,9 +199,9 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable)
[31/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureInMemoryChore.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureInMemoryChore.html b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureInMemoryChore.html index e3b7912..b5ac659 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureInMemoryChore.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureInMemoryChore.html @@ -243,7 +243,7 @@ extends Procedure -acquireLock, addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, doReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, lockedWhenLoading, releaseLock, removeStackIndex, restoreLock, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParent ProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailu re, shouldWaitClientAck, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, tryRunnable, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted +acquireLock, addStackIndex, afterReplay, beforeReplay, bypass, compareTo, completionCleanup, doAcquireLock, doExecute, doReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, lockedWhenLoading, releaseLock, removeStackIndex, restoreLo ck, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, tryRunnable, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureUtil.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureUtil.html b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureUtil.html index 9e56968..5dfd906 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureUtil.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureUtil.html @@ -299,7 +299,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html convertToProcedure -public staticProcedureconvertToProcedure(org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedureproto) +public staticProcedureconvertToProcedure(org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedureproto) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Helper to convert the protobuf procedure. Used by ProcedureStore implementations. @@ -320,7 +320,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html convertToProtoResourceType -public staticorg.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResourceTypeconvertToProtoResourceType(LockedResourceTyperesourceType) +public staticorg.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResourceTypeconvertToProtoResourceType(LockedResourceTyperesourceType) @@ -329,7 +329,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html convertToProtoLockType -public staticorg.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockTypeconvertToProtoLockType(LockTypelockType) +public
[31/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html index 78cd129..17ea7d8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html @@ -51,2276 +51,2305 @@ 043import org.apache.hadoop.hbase.TableName; 044import org.apache.hadoop.hbase.UnknownRegionException; 045import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -046import org.apache.hadoop.hbase.client.MasterSwitchType; -047import org.apache.hadoop.hbase.client.RegionInfo; -048import org.apache.hadoop.hbase.client.RegionInfoBuilder; -049import org.apache.hadoop.hbase.client.TableDescriptor; -050import org.apache.hadoop.hbase.client.TableState; -051import org.apache.hadoop.hbase.client.VersionInfoUtil; -052import org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil; -053import org.apache.hadoop.hbase.coprocessor.MasterCoprocessor; -054import org.apache.hadoop.hbase.errorhandling.ForeignException; -055import org.apache.hadoop.hbase.exceptions.UnknownProtocolException; -056import org.apache.hadoop.hbase.io.hfile.HFile; -057import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -058import org.apache.hadoop.hbase.ipc.PriorityFunction; -059import org.apache.hadoop.hbase.ipc.QosPriority; -060import org.apache.hadoop.hbase.ipc.RpcServer; -061import org.apache.hadoop.hbase.ipc.RpcServer.BlockingServiceAndInterface; -062import org.apache.hadoop.hbase.ipc.RpcServerFactory; -063import org.apache.hadoop.hbase.ipc.RpcServerInterface; -064import org.apache.hadoop.hbase.ipc.ServerRpcController; -065import org.apache.hadoop.hbase.master.assignment.RegionStates; -066import org.apache.hadoop.hbase.master.locking.LockProcedure; -067import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil; -068import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil.NonceProcedureRunnable; -069import org.apache.hadoop.hbase.mob.MobUtils; -070import org.apache.hadoop.hbase.procedure.MasterProcedureManager; -071import org.apache.hadoop.hbase.procedure2.LockType; -072import org.apache.hadoop.hbase.procedure2.LockedResource; -073import org.apache.hadoop.hbase.procedure2.Procedure; -074import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; -075import org.apache.hadoop.hbase.procedure2.ProcedureUtil; -076import org.apache.hadoop.hbase.procedure2.RemoteProcedureException; -077import org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService; -078import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.VisibilityLabelsService; -079import org.apache.hadoop.hbase.quotas.MasterQuotaManager; -080import org.apache.hadoop.hbase.quotas.QuotaObserverChore; -081import org.apache.hadoop.hbase.quotas.QuotaUtil; -082import org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshot; -083import org.apache.hadoop.hbase.regionserver.RSRpcServices; -084import org.apache.hadoop.hbase.regionserver.RpcSchedulerFactory; -085import org.apache.hadoop.hbase.replication.ReplicationException; -086import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -087import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -088import org.apache.hadoop.hbase.security.User; -089import org.apache.hadoop.hbase.security.access.AccessChecker; -090import org.apache.hadoop.hbase.security.access.AccessController; -091import org.apache.hadoop.hbase.security.access.Permission; -092import org.apache.hadoop.hbase.security.visibility.VisibilityController; -093import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; -094import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils; -095import org.apache.hadoop.hbase.util.Bytes; -096import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -097import org.apache.hadoop.hbase.util.ForeignExceptionUtil; -098import org.apache.hadoop.hbase.util.Pair; -099import org.apache.yetus.audience.InterfaceAudience; -100import org.apache.zookeeper.KeeperException; -101import org.slf4j.Logger; -102import org.slf4j.LoggerFactory; -103 -104import org.apache.hbase.thirdparty.com.google.protobuf.RpcController; -105import org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; -106import org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations; -107 -108import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -109import org.apache.hadoop.hbase.shaded.protobuf.ResponseConverter; -110import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -111import
[31/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html b/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html index 4804996..c3cf972 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html @@ -28,500 +28,538 @@ 020import java.util.ArrayList; 021import java.util.Collections; 022import java.util.List; -023 -024import org.apache.hadoop.hbase.Cell; -025import org.apache.hadoop.hbase.CellUtil; -026import org.apache.hadoop.hbase.HConstants; -027import org.apache.hadoop.hbase.PrivateCellUtil; -028import org.apache.yetus.audience.InterfaceAudience; -029import org.apache.hadoop.hbase.exceptions.DeserializationException; -030import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException; -031import org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations; -032import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos; -033import org.apache.hadoop.hbase.util.Bytes; -034 -035/** -036 * Filter to support scan multiple row key ranges. It can construct the row key ranges from the -037 * passed list which can be accessed by each region server. -038 * -039 * HBase is quite efficient when scanning only one small row key range. If user needs to specify -040 * multiple row key ranges in one scan, the typical solutions are: 1. through FilterList which is a -041 * list of row key Filters, 2. using the SQL layer over HBase to join with two table, such as hive, -042 * phoenix etc. However, both solutions are inefficient. Both of them can't utilize the range info -043 * to perform fast forwarding during scan which is quite time consuming. If the number of ranges -044 * are quite big (e.g. millions), join is a proper solution though it is slow. However, there are -045 * cases that user wants to specify a small number of ranges to scan (e.g. lt;1000 ranges). Both -046 * solutions can't provide satisfactory performance in such case. MultiRowRangeFilter is to support -047 * such usec ase (scan multiple row key ranges), which can construct the row key ranges from user -048 * specified list and perform fast-forwarding during scan. Thus, the scan will be quite efficient. -049 */ -050@InterfaceAudience.Public -051public class MultiRowRangeFilter extends FilterBase { -052 -053 private ListRowRange rangeList; -054 -055 private static final int ROW_BEFORE_FIRST_RANGE = -1; -056 private boolean EXCLUSIVE = false; -057 private boolean done = false; -058 private boolean initialized = false; -059 private int index; -060 private RowRange range; -061 private ReturnCode currentReturnCode; -062 -063 /** -064 * @param list A list of codeRowRange/code -065 */ -066 public MultiRowRangeFilter(ListRowRange list) { -067this.rangeList = sortAndMerge(list); -068 } -069 -070 @Override -071 public boolean filterAllRemaining() { -072return done; -073 } -074 -075 public ListRowRange getRowRanges() { -076return this.rangeList; -077 } -078 -079 @Override -080 public boolean filterRowKey(Cell firstRowCell) { -081if (filterAllRemaining()) return true; -082// If it is the first time of running, calculate the current range index for -083// the row key. If index is out of bound which happens when the start row -084// user sets is after the largest stop row of the ranges, stop the scan. -085// If row key is after the current range, find the next range and update index. -086byte[] rowArr = firstRowCell.getRowArray(); -087int length = firstRowCell.getRowLength(); -088int offset = firstRowCell.getRowOffset(); -089if (!initialized -090|| !range.contains(rowArr, offset, length)) { -091 byte[] rowkey = CellUtil.cloneRow(firstRowCell); -092 index = getNextRangeIndex(rowkey); -093 if (index = rangeList.size()) { -094done = true; -095currentReturnCode = ReturnCode.NEXT_ROW; -096return false; -097 } -098 if(index != ROW_BEFORE_FIRST_RANGE) { -099range = rangeList.get(index); -100 } else { -101range = rangeList.get(0); -102 } -103 if (EXCLUSIVE) { -104EXCLUSIVE = false; -105currentReturnCode = ReturnCode.NEXT_ROW; -106return false; -107 } -108 if (!initialized) { -109if(index != ROW_BEFORE_FIRST_RANGE) { -110 currentReturnCode = ReturnCode.INCLUDE; -111} else { -112 currentReturnCode = ReturnCode.SEEK_NEXT_USING_HINT; -113} -114initialized = true; -115 } else { -116if (range.contains(rowArr, offset, length)) { -117
[31/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html index 16913e6..e4622c3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionPipeline.html @@ -80,272 +80,279 @@ 072 } 073 074 public boolean pushHead(MutableSegment segment) { -075ImmutableSegment immutableSegment = SegmentFactory.instance(). -076 createImmutableSegment(segment); -077synchronized (pipeline){ -078 boolean res = addFirst(immutableSegment); -079 readOnlyCopy = new LinkedList(pipeline); -080 return res; -081} -082 } -083 -084 public VersionedSegmentsList getVersionedList() { -085synchronized (pipeline){ -086 return new VersionedSegmentsList(readOnlyCopy, version); -087} -088 } -089 -090 public VersionedSegmentsList getVersionedTail() { -091synchronized (pipeline){ -092 ListImmutableSegment segmentList = new ArrayList(); -093 if(!pipeline.isEmpty()) { -094segmentList.add(0, pipeline.getLast()); -095 } -096 return new VersionedSegmentsList(segmentList, version); -097} -098 } -099 -100 /** -101 * Swaps the versioned list at the tail of the pipeline with a new segment. -102 * Swapping only if there were no changes to the suffix of the list since the version list was -103 * created. -104 * @param versionedList suffix of the pipeline to be replaced can be tail or all the pipeline -105 * @param segment new segment to replace the suffix. Can be null if the suffix just needs to be -106 *removed. -107 * @param closeSuffix whether to close the suffix (to release memory), as part of swapping it out -108 *During index merge op this will be false and for compaction it will be true. -109 * @param updateRegionSize whether to update the region size. Update the region size, -110 * when the pipeline is swapped as part of in-memory-flush and -111 * further merge/compaction. Don't update the region size when the -112 * swap is result of the snapshot (flush-to-disk). -113 * @return true iff swapped tail with new segment -114 */ -115 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="VO_VOLATILE_INCREMENT", -116 justification="Increment is done under a synchronize block so safe") -117 public boolean swap(VersionedSegmentsList versionedList, ImmutableSegment segment, -118 boolean closeSuffix, boolean updateRegionSize) { -119if (versionedList.getVersion() != version) { -120 return false; -121} -122ListImmutableSegment suffix; -123synchronized (pipeline){ -124 if(versionedList.getVersion() != version) { -125return false; -126 } -127 suffix = versionedList.getStoreSegments(); -128 LOG.debug("Swapping pipeline suffix; before={}, new segment={}", -129 versionedList.getStoreSegments().size(), segment); -130 swapSuffix(suffix, segment, closeSuffix); -131 readOnlyCopy = new LinkedList(pipeline); -132 version++; -133} -134if (updateRegionSize region != null) { -135 // update the global memstore size counter -136 long suffixDataSize = getSegmentsKeySize(suffix); -137 long newDataSize = 0; -138 if(segment != null) { -139newDataSize = segment.getDataSize(); -140 } -141 long dataSizeDelta = suffixDataSize - newDataSize; -142 long suffixHeapSize = getSegmentsHeapSize(suffix); -143 long suffixOffHeapSize = getSegmentsOffHeapSize(suffix); -144 long newHeapSize = 0; -145 long newOffHeapSize = 0; -146 if(segment != null) { -147newHeapSize = segment.getHeapSize(); -148newOffHeapSize = segment.getOffHeapSize(); -149 } -150 long offHeapSizeDelta = suffixOffHeapSize - newOffHeapSize; -151 long heapSizeDelta = suffixHeapSize - newHeapSize; -152 region.addMemStoreSize(-dataSizeDelta, -heapSizeDelta, -offHeapSizeDelta); -153 LOG.debug("Suffix data size={}, new segment data size={}, " + "suffix heap size={}," -154 + "new segment heap size={}" + "suffix off heap size={}," -155 + "new segment off heap size={}", suffixDataSize, newDataSize, suffixHeapSize, -156 newHeapSize, suffixOffHeapSize, newOffHeapSize); -157} -158return true; -159 } -160 -161 private static long getSegmentsHeapSize(List? extends Segment list) { -162long res = 0; -163for (Segment segment : list) { -164 res += segment.getHeapSize(); -165} -166
[31/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStateNode.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStateNode.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStateNode.html new file mode 100644 index 000..3525c1e --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStateNode.html @@ -0,0 +1,952 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +RegionStateNode (Apache HBase 3.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.master.assignment +Class RegionStateNode + + + +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.master.assignment.RegionStateNode + + + + + + + +All Implemented Interfaces: +https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableRegionStateNode + + + +@InterfaceAudience.Private +public class RegionStateNode +extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +implements https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableRegionStateNode +Current Region State. Most fields are synchronized with meta region, i.e, we will update meta + immediately after we modify this RegionStateNode, and usually under the lock. The only exception + is lastHost, which should not be used for critical condition. + + Typically, the only way to modify this class is through TransitRegionStateProcedure, and + we will record the TRSP along with this RegionStateNode to make sure that there could at most one + TRSP. For other operations, such as SCP, we will first get the lock, and then try to schedule a + TRSP. If there is already one, then the solution will be different: + + For SCP, we will update the region state in meta to tell the TRSP to retry. + For DisableTableProcedure, as we have the xlock, we can make sure that the TRSP has not been + executed yet, so just unset it and attach a new one. The original one will quit immediately when + executing. + For split/merge, we will fail immediately as there is no actual operations yet so no + harm. + For EnableTableProcedure/TruncateTableProcedure, we can make sure that there will be no TRSP + attached with the RSNs. + For other procedures, you'd better use ReopenTableRegionsProcedure. The RTRP will take care + of lots of corner cases when reopening regions. + + + Several fields are declared with volatile, which means you are free to get it without + lock, but usually you should not use these fields without locking for critical condition, as it + will be easily to introduce inconsistency. For example, you are free to dump the status and show + it on web without locking, but if you want to change the state of the RegionStateNode by checking + the current state, you'd better have the lock... + + + + + + + + + + + +Nested Class Summary + +Nested Classes + +Modifier and Type +Class and Description + + +private static class +RegionStateNode.AssignmentProcedureEvent + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +private ProcedureEvent? +event + + +private ServerName +lastHost + + +private long +lastUpdate +Updated whenever a call to setRegionLocation(ServerName) or + #setState(State, State...). + + + +(package private)
[31/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellReducer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellReducer.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellReducer.html index 39170f0..7859ebc 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellReducer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellReducer.html @@ -230,564 +230,567 @@ 222 } 223} 224 } catch (InterruptedException e) { -225e.printStackTrace(); -226 } -227} -228 -229@Override -230public void setup(Context context) throws IOException { -231 cfRenameMap = createCfRenameMap(context.getConfiguration()); -232 filter = instantiateFilter(context.getConfiguration()); -233 int reduceNum = context.getNumReduceTasks(); -234 Configuration conf = context.getConfiguration(); -235 TableName tableName = TableName.valueOf(context.getConfiguration().get(TABLE_NAME)); -236 try (Connection conn = ConnectionFactory.createConnection(conf); -237 RegionLocator regionLocator = conn.getRegionLocator(tableName)) { -238byte[][] startKeys = regionLocator.getStartKeys(); -239if (startKeys.length != reduceNum) { -240 throw new IOException("Region split after job initialization"); -241} -242CellWritableComparable[] startKeyWraps = -243new CellWritableComparable[startKeys.length - 1]; -244for (int i = 1; i startKeys.length; ++i) { -245 startKeyWraps[i - 1] = -246 new CellWritableComparable(KeyValueUtil.createFirstOnRow(startKeys[i])); -247} -248 CellWritableComparablePartitioner.START_KEYS = startKeyWraps; -249 } -250} -251 } -252 -253 /** -254 * A mapper that just writes out KeyValues. -255 */ -256 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="EQ_COMPARETO_USE_OBJECT_EQUALS", -257 justification="Writables are going away and this has been this way forever") -258 public static class CellImporter extends TableMapperImmutableBytesWritable, Cell { -259private Mapbyte[], byte[] cfRenameMap; -260private Filter filter; -261private static final Logger LOG = LoggerFactory.getLogger(CellImporter.class); -262 -263/** -264 * @param row The current table row key. -265 * @param value The columns. -266 * @param context The current context. -267 * @throws IOException When something is broken with the data. -268 */ -269@Override -270public void map(ImmutableBytesWritable row, Result value, -271 Context context) -272throws IOException { -273 try { -274if (LOG.isTraceEnabled()) { -275 LOG.trace("Considering the row." -276 + Bytes.toString(row.get(), row.getOffset(), row.getLength())); -277} -278if (filter == null -279|| !filter.filterRowKey(PrivateCellUtil.createFirstOnRow(row.get(), row.getOffset(), -280(short) row.getLength( { -281 for (Cell kv : value.rawCells()) { -282kv = filterKv(filter, kv); -283// skip if we filtered it out -284if (kv == null) continue; -285context.write(row, new MapReduceExtendedCell(convertKv(kv, cfRenameMap))); -286 } -287} -288 } catch (InterruptedException e) { -289e.printStackTrace(); -290 } -291} -292 -293@Override -294public void setup(Context context) { -295 cfRenameMap = createCfRenameMap(context.getConfiguration()); -296 filter = instantiateFilter(context.getConfiguration()); -297} -298 } -299 -300 /** -301 * Write table content out to files in hdfs. -302 */ -303 public static class Importer extends TableMapperImmutableBytesWritable, Mutation { -304private Mapbyte[], byte[] cfRenameMap; -305private ListUUID clusterIds; -306private Filter filter; -307private Durability durability; -308 -309/** -310 * @param row The current table row key. -311 * @param value The columns. -312 * @param context The current context. -313 * @throws IOException When something is broken with the data. -314 */ -315@Override -316public void map(ImmutableBytesWritable row, Result value, -317 Context context) -318throws IOException { -319 try { -320writeResult(row, value, context); -321 } catch (InterruptedException e) { -322e.printStackTrace(); -323 } -324} -325 -326private void writeResult(ImmutableBytesWritable key, Result result, Context context) -327throws IOException, InterruptedException { -328 Put put = null; -329 Delete delete = null; -330 if (LOG.isTraceEnabled()) {
[31/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/CacheableDeserializerIdManager.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/CacheableDeserializerIdManager.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/CacheableDeserializerIdManager.html index e50f682..7d5287a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/CacheableDeserializerIdManager.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/CacheableDeserializerIdManager.html @@ -33,37 +33,55 @@ 025import org.apache.yetus.audience.InterfaceAudience; 026 027/** -028 * This class is used to manage the identifiers for -029 * {@link CacheableDeserializer} -030 */ -031@InterfaceAudience.Private -032public class CacheableDeserializerIdManager { -033 private static final MapInteger, CacheableDeserializerCacheable registeredDeserializers = new HashMap(); -034 private static final AtomicInteger identifier = new AtomicInteger(0); -035 -036 /** -037 * Register the given cacheable deserializer and generate an unique identifier -038 * id for it -039 * @param cd -040 * @return the identifier of given cacheable deserializer -041 */ -042 public static int registerDeserializer(CacheableDeserializerCacheable cd) { -043int idx = identifier.incrementAndGet(); -044synchronized (registeredDeserializers) { -045 registeredDeserializers.put(idx, cd); -046} -047return idx; -048 } -049 -050 /** -051 * Get the cacheable deserializer as the given identifier Id -052 * @param id -053 * @return CacheableDeserializer -054 */ -055 public static CacheableDeserializerCacheable getDeserializer(int id) { -056return registeredDeserializers.get(id); -057 } -058} +028 * This class is used to manage the identifiers for {@link CacheableDeserializer}. +029 * All deserializers are registered with this Manager via the +030 * {@link #registerDeserializer(CacheableDeserializer)}}. On registration, we return an +031 * int *identifier* for this deserializer. The int identifier is passed to +032 * {@link #getDeserializer(int)}} to obtain the registered deserializer instance. +033 */ +034@InterfaceAudience.Private +035public class CacheableDeserializerIdManager { +036 private static final MapInteger, CacheableDeserializerCacheable registeredDeserializers = new HashMap(); +037 private static final AtomicInteger identifier = new AtomicInteger(0); +038 +039 /** +040 * Register the given {@link Cacheable} -- usually an hfileblock instance, these implement +041 * the Cacheable Interface -- deserializer and generate an unique identifier id for it and return +042 * this as our result. +043 * @return the identifier of given cacheable deserializer +044 * @see #getDeserializer(int) +045 */ +046 public static int registerDeserializer(CacheableDeserializerCacheable cd) { +047int idx = identifier.incrementAndGet(); +048synchronized (registeredDeserializers) { +049 registeredDeserializers.put(idx, cd); +050} +051return idx; +052 } +053 +054 /** +055 * Get the cacheable deserializer registered at the given identifier Id. +056 * @see #registerDeserializer(CacheableDeserializer) +057 */ +058 public static CacheableDeserializerCacheable getDeserializer(int id) { +059return registeredDeserializers.get(id); +060 } +061 +062 /** +063 * Snapshot a map of the current identifiers to class names for reconstruction on reading out +064 * of a file. +065 */ +066 public static MapInteger,String save() { +067MapInteger, String snapshot = new HashMap(); +068synchronized (registeredDeserializers) { +069 for (Map.EntryInteger, CacheableDeserializerCacheable entry : +070 registeredDeserializers.entrySet()) { +071snapshot.put(entry.getKey(), entry.getValue().getClass().getName()); +072 } +073} +074return snapshot; +075 } +076}
[31/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/procedure/InitMetaProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/InitMetaProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/InitMetaProcedure.html index 2f903de..a6743fd 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/InitMetaProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/InitMetaProcedure.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -295,6 +295,14 @@ extends +protected boolean +waitInitialized(MasterProcedureEnvenv) +The Procedure.doAcquireLock(Object, ProcedureStore) will be split into two steps, first, it will + call us to determine whether we need to wait for initialization, second, it will call + Procedure.acquireLock(Object) to actually handle the lock for this procedure. + + @@ -315,7 +323,7 @@ extends Procedure -addStackIndex, afterReplay, beforeReplay, compareTo, doAcquireLock, doExecute, doReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime< /a>, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted +addStackIndex, afterReplay, beforeReplay, compareTo, doExecute, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, get ProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted @@ -437,35 +445,60 @@ extends + + + + +waitInitialized +protectedbooleanwaitInitialized(MasterProcedureEnvenv) +Description copied from class:Procedure +The Procedure.doAcquireLock(Object, ProcedureStore) will be split into two steps, first, it will + call us to determine whether we need to wait for initialization, second, it will call + Procedure.acquireLock(Object) to actually handle the lock for this procedure. + + This is because that when master restarts, we need to restore the lock state for all the + procedures to not break the semantic if Procedure.holdLock(Object) is true. But the + ProcedureExecutor will be started before the master finish initialization(as it is part + of the initialization!), so we need to split the code into two steps, and when restore, we just + restore the lock part and ignore the waitInitialized part. Otherwise there will be dead lock. + +Overrides: +waitInitializedin classAbstractStateMachineTableProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.InitMetaState +Returns: +true means we need to wait until the environment has been initialized, otherwise true. + + + acquireLock -protectedProcedure.LockStateacquireLock(MasterProcedureEnvenv) +protectedProcedure.LockStateacquireLock(MasterProcedureEnvenv) Description copied from class:Procedure -The user should override this method if they need a lock on an Entity. - A lock can be anything, and it is up to the implementor.
[31/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html index 5f41fe7..c8158b5 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html @@ -524,636 +524,639 @@ 516} 517 } 518 -519 ArrayListRegionStateNode getTableRegionStateNodes(final TableName tableName) { -520final ArrayListRegionStateNode regions = new ArrayListRegionStateNode(); -521for (RegionStateNode node: regionsMap.tailMap(tableName.getName()).values()) { -522 if (!node.getTable().equals(tableName)) break; -523 regions.add(node); -524} -525return regions; -526 } -527 -528 ArrayListRegionState getTableRegionStates(final TableName tableName) { -529final ArrayListRegionState regions = new ArrayListRegionState(); -530for (RegionStateNode node: regionsMap.tailMap(tableName.getName()).values()) { -531 if (!node.getTable().equals(tableName)) break; -532 regions.add(node.toRegionState()); -533} -534return regions; -535 } -536 -537 ArrayListRegionInfo getTableRegionsInfo(final TableName tableName) { -538final ArrayListRegionInfo regions = new ArrayListRegionInfo(); -539for (RegionStateNode node: regionsMap.tailMap(tableName.getName()).values()) { -540 if (!node.getTable().equals(tableName)) break; -541 regions.add(node.getRegionInfo()); -542} -543return regions; -544 } -545 -546 CollectionRegionStateNode getRegionStateNodes() { -547return regionsMap.values(); +519 public void deleteRegions(final ListRegionInfo regionInfos) { +520 regionInfos.forEach(this::deleteRegion); +521 } +522 +523 ArrayListRegionStateNode getTableRegionStateNodes(final TableName tableName) { +524final ArrayListRegionStateNode regions = new ArrayListRegionStateNode(); +525for (RegionStateNode node: regionsMap.tailMap(tableName.getName()).values()) { +526 if (!node.getTable().equals(tableName)) break; +527 regions.add(node); +528} +529return regions; +530 } +531 +532 ArrayListRegionState getTableRegionStates(final TableName tableName) { +533final ArrayListRegionState regions = new ArrayListRegionState(); +534for (RegionStateNode node: regionsMap.tailMap(tableName.getName()).values()) { +535 if (!node.getTable().equals(tableName)) break; +536 regions.add(node.toRegionState()); +537} +538return regions; +539 } +540 +541 ArrayListRegionInfo getTableRegionsInfo(final TableName tableName) { +542final ArrayListRegionInfo regions = new ArrayListRegionInfo(); +543for (RegionStateNode node: regionsMap.tailMap(tableName.getName()).values()) { +544 if (!node.getTable().equals(tableName)) break; +545 regions.add(node.getRegionInfo()); +546} +547return regions; 548 } 549 -550 public ArrayListRegionState getRegionStates() { -551final ArrayListRegionState regions = new ArrayListRegionState(regionsMap.size()); -552for (RegionStateNode node: regionsMap.values()) { -553 regions.add(node.toRegionState()); -554} -555return regions; -556 } -557 -558 // == -559 // RegionState helpers -560 // == -561 public RegionState getRegionState(final RegionInfo regionInfo) { -562RegionStateNode regionStateNode = getRegionStateNode(regionInfo); -563return regionStateNode == null ? null : regionStateNode.toRegionState(); -564 } -565 -566 public RegionState getRegionState(final String encodedRegionName) { -567// TODO: Need a map encodedName, ... but it is just dispatch merge... -568for (RegionStateNode node: regionsMap.values()) { -569 if (node.getRegionInfo().getEncodedName().equals(encodedRegionName)) { -570return node.toRegionState(); -571 } -572} -573return null; -574 } -575 -576 // -577 // TODO: helpers -578 // -579 public boolean hasTableRegionStates(final TableName tableName) { -580// TODO -581return !getTableRegionStates(tableName).isEmpty(); -582 } -583 -584 /** -585 * @return Return online regions of table; does not include OFFLINE or SPLITTING regions. -586 */ -587 public ListRegionInfo getRegionsOfTable(final TableName table) { -588return getRegionsOfTable(table, false); -589 } -590 -591 private
[31/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/HRegionInfo.html -- diff --git a/apidocs/org/apache/hadoop/hbase/HRegionInfo.html b/apidocs/org/apache/hadoop/hbase/HRegionInfo.html index cb3cceb..1655e52 100644 --- a/apidocs/org/apache/hadoop/hbase/HRegionInfo.html +++ b/apidocs/org/apache/hadoop/hbase/HRegionInfo.html @@ -1,6 +1,6 @@ http://www.w3.org/TR/html4/loose.dtd;> - + @@ -19,45 +19,45 @@ } //--> var methods = {"i0":41,"i1":42,"i2":42,"i3":42,"i4":41,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":42,"i12":42,"i13":42,"i14":42,"i15":41,"i16":41,"i17":42,"i18":41,"i19":42,"i20":41,"i21":42,"i22":41,"i23":42,"i24":42,"i25":42,"i26":42,"i27":42,"i28":42,"i29":41,"i30":41,"i31":41,"i32":41,"i33":41,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42}; -var tabs = {65535:["t0","æææ¹æ³"],1:["t1","éææ¹æ³"],2:["t2","å®ä¾æ¹æ³"],8:["t4","å ·ä½æ¹æ³"],32:["t6","å·²è¿æ¶çæ¹æ³"]}; +var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; var activeTableTab = "activeTableTab"; -æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã +JavaScript is disabled on your browser. -è·³è¿å¯¼èªé¾æ¥ +Skip navigation links - -æ¦è§ -ç¨åºå -ç±» -ä½¿ç¨ -æ -å·²è¿æ¶ -ç´¢å¼ -å¸®å© + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help -ä¸ä¸ä¸ªç±» -ä¸ä¸ä¸ªç±» +PrevClass +NextClass -æ¡æ¶ -æ æ¡æ¶ +Frames +NoFrames -ææç±» +AllClasses
[31/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/HConstants.html -- diff --git a/apidocs/org/apache/hadoop/hbase/HConstants.html b/apidocs/org/apache/hadoop/hbase/HConstants.html index 51e6d1a..abd788f 100644 --- a/apidocs/org/apache/hadoop/hbase/HConstants.html +++ b/apidocs/org/apache/hadoop/hbase/HConstants.html @@ -1,6 +1,6 @@ http://www.w3.org/TR/html4/loose.dtd;> - + @@ -20,38 +20,38 @@ //--> -JavaScript is disabled on your browser. +æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã -Skip navigation links +è·³è¿å¯¼èªé¾æ¥ - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help + +æ¦è§ +ç¨åºå +ç±» +ä½¿ç¨ +æ +å·²è¿æ¶ +ç´¢å¼ +å¸®å© -PrevClass -NextClass +ä¸ä¸ä¸ªç±» +ä¸ä¸ä¸ªç±» -Frames -NoFrames +æ¡æ¶ +æ æ¡æ¶ -AllClasses +ææç±»
[31/51] [partial] hbase-site git commit: Published site at 0f23784182ab88649de340d75804e0ff20dcd0fc.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html index db9b75d..8c494308 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html @@ -3142,7 +3142,7 @@ implements closeLock -private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object closeLock +private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object closeLock @@ -3151,7 +3151,7 @@ implements MEMSTORE_PERIODIC_FLUSH_INTERVAL -public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MEMSTORE_PERIODIC_FLUSH_INTERVAL +public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MEMSTORE_PERIODIC_FLUSH_INTERVAL Conf key for the periodic flush interval See Also: @@ -3165,7 +3165,7 @@ implements DEFAULT_CACHE_FLUSH_INTERVAL -public static finalint DEFAULT_CACHE_FLUSH_INTERVAL +public static finalint DEFAULT_CACHE_FLUSH_INTERVAL Default interval for the memstore flush See Also: @@ -3179,7 +3179,7 @@ implements SYSTEM_CACHE_FLUSH_INTERVAL -public static finalint SYSTEM_CACHE_FLUSH_INTERVAL +public static finalint SYSTEM_CACHE_FLUSH_INTERVAL Default interval for System tables memstore flush See Also: @@ -3193,7 +3193,7 @@ implements MEMSTORE_FLUSH_PER_CHANGES -public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MEMSTORE_FLUSH_PER_CHANGES +public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MEMSTORE_FLUSH_PER_CHANGES Conf key to force a flush if there are already enough changes for one region in memstore See Also: @@ -3207,7 +3207,7 @@ implements DEFAULT_FLUSH_PER_CHANGES -public static finallong DEFAULT_FLUSH_PER_CHANGES +public static finallong DEFAULT_FLUSH_PER_CHANGES See Also: Constant Field Values @@ -3220,7 +3220,7 @@ implements MAX_FLUSH_PER_CHANGES -public static finallong MAX_FLUSH_PER_CHANGES +public static finallong MAX_FLUSH_PER_CHANGES The following MAX_FLUSH_PER_CHANGES is large enough because each KeyValue has 20+ bytes overhead. Therefore, even 1G empty KVs occupy at least 20GB memstore size for a single region @@ -3235,7 +3235,7 @@ implements FOR_UNIT_TESTS_ONLY -private static finalbyte[] FOR_UNIT_TESTS_ONLY +private static finalbyte[] FOR_UNIT_TESTS_ONLY Row needed by below method. @@ -3245,7 +3245,7 @@ implements FIXED_OVERHEAD -public static finallong FIXED_OVERHEAD +public static finallong FIXED_OVERHEAD @@ -3254,7 +3254,7 @@ implements DEEP_OVERHEAD -public static finallong DEEP_OVERHEAD +public static finallong DEEP_OVERHEAD @@ -3263,7 +3263,7 @@ implements MOCKED_LIST -private static finalhttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCell MOCKED_LIST +private static finalhttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCell MOCKED_LIST A mocked list implementation - discards all updates. @@ -3434,7 +3434,7 @@ publiclong initializeStores -privatelonginitializeStores(CancelableProgressablereporter, +privatelonginitializeStores(CancelableProgressablereporter, MonitoredTaskstatus) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Open all Stores. @@ -3455,7 +3455,7 @@ publiclong initializeWarmup -privatevoidinitializeWarmup(CancelableProgressablereporter) +privatevoidinitializeWarmup(CancelableProgressablereporter) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -3469,7 +3469,7 @@ publiclong getStoreFiles -privatehttps://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true; title="class or interface in java.util">NavigableMapbyte[],https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.PathgetStoreFiles()
[31/51] [partial] hbase-site git commit: Published site at 85b41f36e01214b6485c9352875c84ebf877dab3.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionState.State.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionState.State.html b/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionState.State.html index 69a3eee..c314dc4 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionState.State.html +++ b/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionState.State.html @@ -407,10 +407,9 @@ the order they are declared. private void -RegionStateStore.updateUserRegionLocation(RegionInforegionInfo, +RegionStateStore.updateUserRegionLocation(RegionInforegionInfo, RegionState.Statestate, ServerNameregionLocation, -ServerNamelastHost, longopenSeqNum, longpid) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/master/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html index b1a7941..75dcd36 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html @@ -348,11 +348,11 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus -org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective -org.apache.hadoop.hbase.master.RegionState.State org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode +org.apache.hadoop.hbase.master.RegionState.State +org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html index de7b6db..e392df8 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html @@ -217,9 +217,9 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType -org.apache.hadoop.hbase.master.procedure.MetaProcedureInterface.MetaOperationType -org.apache.hadoop.hbase.master.procedure.PeerProcedureInterface.PeerOperationType org.apache.hadoop.hbase.master.procedure.ServerProcedureInterface.ServerOperationType +org.apache.hadoop.hbase.master.procedure.PeerProcedureInterface.PeerOperationType +org.apache.hadoop.hbase.master.procedure.MetaProcedureInterface.MetaOperationType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/metrics/class-use/Metric.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/metrics/class-use/Metric.html b/devapidocs/org/apache/hadoop/hbase/metrics/class-use/Metric.html index 4c77c55..385955b 100644 --- a/devapidocs/org/apache/hadoop/hbase/metrics/class-use/Metric.html +++ b/devapidocs/org/apache/hadoop/hbase/metrics/class-use/Metric.html @@ -83,12 +83,18 @@ +org.apache.hadoop.hbase.coprocessor + +Table of Contents + + + org.apache.hadoop.hbase.metrics Metrics API for HBase. - + org.apache.hadoop.hbase.metrics.impl Implementation of the HBase Metrics framework. @@ -99,6 +105,24 @@ + + + +Uses of Metric in org.apache.hadoop.hbase.coprocessor + +Fields in org.apache.hadoop.hbase.coprocessor with type parameters of type Metric + +Modifier and Type +Field and Description + + + +private
[31/51] [partial] hbase-site git commit: Published site at 6198e1fc7dfa85c3bc6b2855f9a5fb5f4b2354ff.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html b/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html index 6b3b8d1..48f2d17 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html +++ b/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -class AsyncHBaseAdmin +class AsyncHBaseAdmin extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements AsyncAdmin Just a wrapper of RawAsyncHBaseAdmin. The difference is that users need to provide a @@ -976,39 +976,46 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +transitReplicationPeerSyncReplicationState(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringpeerId, + SyncReplicationStateclusterState) +Transit current cluster to a new state in a synchronous replication peer. + + + +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void truncateTable(TableNametableName, booleanpreserveSplits) Truncate a table. - + https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void unassign(byte[]regionName, booleanforcible) Unassign a region
[31/51] [partial] hbase-site git commit: Published site at 14087cc919da9f2e0b1a68f701f6365ad9d1d71f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html index 490dd78..943c050 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -419,170 +419,166 @@ implements RegionInforegionToMergeB) -MoveRegionProcedure +private MoveRegionProcedure createMoveRegionProcedure(RegionPlanplan) -MoveRegionProcedure[] -createReopenProcedures(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true; title="class or interface in java.util">CollectionRegionInforegionInfo) - - AssignProcedure[] createRoundRobinAssignProcedures(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfohris) Create round-robin assigns. - + SplitTableRegionProcedure createSplitProcedure(RegionInforegionToSplit, byte[]splitKey) - + (package private) UnassignProcedure createUnassignProcedure(RegionInforegionInfo, ServerNamedestinationServer, booleanforce) - + (package private) UnassignProcedure createUnassignProcedure(RegionInforegionInfo, ServerNamedestinationServer, booleanforce, booleanremoveAfterUnassigning) - + (package private) UnassignProcedure[] createUnassignProcedures(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true; title="class or interface in java.util">CollectionRegionStates.RegionStateNodenodes) - + UnassignProcedure[] createUnassignProcedures(TableNametableName) Called by things like DisableTableProcedure to get a list of UnassignProcedure to unassign the regions of the table. - + void deleteTable(TableNametableName) Delete the region states. - + https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo getAssignedRegions() - + (package private) int getAssignMaxAttempts() - + MetricsAssignmentManager getAssignmentManagerMetrics() - + private LoadBalancer getBalancer() - + org.apache.hadoop.conf.Configuration getConfiguration() - + https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName
[31/51] [partial] hbase-site git commit: Published site at 72784c2d836a4b977667449d3adec5e8d15453f5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html index b6e7636..592c2cc 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html @@ -356,3901 +356,3924 @@ 348 public FutureVoid modifyTableAsync(TableDescriptor td) throws IOException { 349ModifyTableResponse response = executeCallable( 350 new MasterCallableModifyTableResponse(getConnection(), getRpcControllerFactory()) { -351@Override -352protected ModifyTableResponse rpcCall() throws Exception { -353 setPriority(td.getTableName()); -354 ModifyTableRequest request = RequestConverter.buildModifyTableRequest( -355td.getTableName(), td, ng.getNonceGroup(), ng.newNonce()); -356 return master.modifyTable(getRpcController(), request); -357} -358 }); -359return new ModifyTableFuture(this, td.getTableName(), response); -360 } -361 -362 @Override -363 public ListTableDescriptor listTableDescriptorsByNamespace(byte[] name) throws IOException { -364return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -365getRpcControllerFactory()) { -366 @Override -367 protected ListTableDescriptor rpcCall() throws Exception { -368return master.listTableDescriptorsByNamespace(getRpcController(), -369 ListTableDescriptorsByNamespaceRequest.newBuilder() -370 .setNamespaceName(Bytes.toString(name)).build()) -371.getTableSchemaList() -372.stream() -373 .map(ProtobufUtil::toTableDescriptor) -374 .collect(Collectors.toList()); -375 } -376}); -377 } -378 -379 @Override -380 public ListTableDescriptor listTableDescriptors(ListTableName tableNames) throws IOException { -381return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -382getRpcControllerFactory()) { -383 @Override -384 protected ListTableDescriptor rpcCall() throws Exception { -385GetTableDescriptorsRequest req = -386 RequestConverter.buildGetTableDescriptorsRequest(tableNames); -387 return ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(), -388 req)); -389 } -390}); -391 } -392 -393 @Override -394 public ListRegionInfo getRegions(final ServerName sn) throws IOException { -395AdminService.BlockingInterface admin = this.connection.getAdmin(sn); -396// TODO: There is no timeout on this controller. Set one! -397HBaseRpcController controller = rpcControllerFactory.newController(); -398return ProtobufUtil.getOnlineRegions(controller, admin); -399 } -400 -401 @Override -402 public ListRegionInfo getRegions(TableName tableName) throws IOException { -403if (TableName.isMetaTableName(tableName)) { -404 return Arrays.asList(RegionInfoBuilder.FIRST_META_REGIONINFO); -405} else { -406 return MetaTableAccessor.getTableRegions(connection, tableName, true); -407} -408 } -409 -410 private static class AbortProcedureFuture extends ProcedureFutureBoolean { -411private boolean isAbortInProgress; -412 -413public AbortProcedureFuture( -414final HBaseAdmin admin, -415final Long procId, -416final Boolean abortProcResponse) { -417 super(admin, procId); -418 this.isAbortInProgress = abortProcResponse; -419} -420 -421@Override -422public Boolean get(long timeout, TimeUnit unit) -423throws InterruptedException, ExecutionException, TimeoutException { -424 if (!this.isAbortInProgress) { -425return false; -426 } -427 super.get(timeout, unit); -428 return true; -429} -430 } -431 -432 /** @return Connection used by this object. */ -433 @Override -434 public Connection getConnection() { -435return connection; -436 } -437 -438 @Override -439 public boolean tableExists(final TableName tableName) throws IOException { -440return executeCallable(new RpcRetryingCallableBoolean() { -441 @Override -442 protected Boolean rpcCall(int callTimeout) throws Exception { -443return MetaTableAccessor.tableExists(connection, tableName); -444 } -445}); -446 } -447 -448 @Override -449 public HTableDescriptor[] listTables() throws IOException { -450return listTables((Pattern)null, false); -451 } -452 -453 @Override -454 public HTableDescriptor[] listTables(Pattern pattern) throws
[31/51] [partial] hbase-site git commit: Published site at 9101fc246f86445006bfbcdfda5cc495016dc280.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.MetaOperationType.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.MetaOperationType.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.MetaOperationType.html index 376bec0..5e09558 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.MetaOperationType.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.MetaOperationType.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -public static enum MetaProcedureInterface.MetaOperationType +public static enum MetaProcedureInterface.MetaOperationType extends https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumMetaProcedureInterface.MetaOperationType @@ -207,7 +207,7 @@ the order they are declared. RECOVER -public static finalMetaProcedureInterface.MetaOperationType RECOVER +public static finalMetaProcedureInterface.MetaOperationType RECOVER @@ -224,7 +224,7 @@ the order they are declared. values -public staticMetaProcedureInterface.MetaOperationType[]values() +public staticMetaProcedureInterface.MetaOperationType[]values() Returns an array containing the constants of this enum type, in the order they are declared. This method may be used to iterate over the constants as follows: @@ -244,7 +244,7 @@ for (MetaProcedureInterface.MetaOperationType c : MetaProcedureInterface.MetaOpe valueOf -public staticMetaProcedureInterface.MetaOperationTypevalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) +public staticMetaProcedureInterface.MetaOperationTypevalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Returns the enum constant of this type with the specified name. The string must match exactly an identifier used to declare an enum constant in this type. (Extraneous whitespace characters are http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.html index 46d47b9..c9e23ab 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaProcedureInterface.html @@ -18,8 +18,8 @@ catch(err) { } //--> -var methods = {"i0":18}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],16:["t5","Default Methods"]}; +var methods = {"i0":50}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],16:["t5","Default Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; @@ -104,9 +104,14 @@ var activeTableTab = "activeTableTab"; RecoverMetaProcedure +Deprecated. +only used for RecoverMetaProcedure. Should be removed along with + RecoverMetaProcedure. + -@InterfaceAudience.Private -public interface MetaProcedureInterface +https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true; title="class or interface in java.lang">@Deprecated + @InterfaceAudience.Private +public interface MetaProcedureInterface @@ -127,7 +132,9 @@ public interface static class -MetaProcedureInterface.MetaOperationType +MetaProcedureInterface.MetaOperationType +Deprecated. + @@ -139,14 +146,16 @@ public interface -All MethodsInstance MethodsDefault Methods +All MethodsInstance MethodsDefault MethodsDeprecated Methods Modifier and Type Method and Description default MetaProcedureInterface.MetaOperationType -getMetaOperationType() +getMetaOperationType() +Deprecated. + @@ -169,7 +178,8 @@ public interface getMetaOperationType -defaultMetaProcedureInterface.MetaOperationTypegetMetaOperationType() +defaultMetaProcedureInterface.MetaOperationTypegetMetaOperationType() +Deprecated. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaQueue.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaQueue.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaQueue.html index 7b0ca22..2beef2b 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/MetaQueue.html +++
[31/51] [partial] hbase-site git commit: Published site at 0b28155d274910b4e667b949d51f78809a1eff0b.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html index f236300..513d2ad 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html @@ -26,1048 +26,1115 @@ 018 019package org.apache.hadoop.hbase.backup.impl; 020 -021import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH; -022import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC; -023import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG; -024import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC; -025import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH; -026import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC; -027import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER; -028import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC; -029import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET; -030import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC; -031import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC; -032import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE; -033import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC; -034import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC; -035import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS; -036import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC; -037import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME; -038import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC; -039 -040import java.io.IOException; -041import java.net.URI; -042import java.util.List; +021import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BACKUP_LIST_DESC; +022import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH; +023import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC; +024import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG; +025import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC; +026import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP; +027import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP_DESC; +028import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_LIST; +029import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH; +030import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC; +031import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER; +032import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC; +033import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET; +034import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC; +035import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC; +036import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE; +037import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC; +038import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC; +039import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS; +040import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC; +041import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME; +042import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC; 043 -044import org.apache.commons.lang3.StringUtils; -045import org.apache.hadoop.conf.Configuration; -046import org.apache.hadoop.conf.Configured; -047import org.apache.hadoop.fs.FileSystem; -048import org.apache.hadoop.fs.Path; -049import org.apache.hadoop.hbase.HBaseConfiguration; -050import org.apache.hadoop.hbase.TableName; -051import org.apache.hadoop.hbase.backup.BackupAdmin; -052import org.apache.hadoop.hbase.backup.BackupInfo; -053import
[31/51] [partial] hbase-site git commit: Published site at 7d3750bd9fc9747623549c242cc4171e224b3eaf.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html index aa48364..9549aa5 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html @@ -2830,843 +2830,858 @@ 2822 * @return true if master is in maintenanceMode 2823 */ 2824 @Override -2825 public boolean isInMaintenanceMode() { -2826return maintenanceModeTracker.isInMaintenanceMode(); -2827 } -2828 -2829 @VisibleForTesting -2830 public void setInitialized(boolean isInitialized) { -2831 procedureExecutor.getEnvironment().setEventReady(initialized, isInitialized); -2832 } -2833 -2834 @Override -2835 public ProcedureEvent? getInitializedEvent() { -2836return initialized; -2837 } -2838 -2839 /** -2840 * ServerCrashProcessingEnabled is set false before completing assignMeta to prevent processing -2841 * of crashed servers. -2842 * @return true if assignMeta has completed; -2843 */ -2844 @Override -2845 public boolean isServerCrashProcessingEnabled() { -2846return serverCrashProcessingEnabled.isReady(); -2847 } -2848 -2849 @VisibleForTesting -2850 public void setServerCrashProcessingEnabled(final boolean b) { -2851 procedureExecutor.getEnvironment().setEventReady(serverCrashProcessingEnabled, b); -2852 } -2853 -2854 public ProcedureEvent? getServerCrashProcessingEnabledEvent() { -2855return serverCrashProcessingEnabled; -2856 } -2857 -2858 /** -2859 * Compute the average load across all region servers. -2860 * Currently, this uses a very naive computation - just uses the number of -2861 * regions being served, ignoring stats about number of requests. -2862 * @return the average load -2863 */ -2864 public double getAverageLoad() { -2865if (this.assignmentManager == null) { -2866 return 0; -2867} -2868 -2869RegionStates regionStates = this.assignmentManager.getRegionStates(); -2870if (regionStates == null) { -2871 return 0; -2872} -2873return regionStates.getAverageLoad(); -2874 } -2875 -2876 /* -2877 * @return the count of region split plans executed -2878 */ -2879 public long getSplitPlanCount() { -2880return splitPlanCount; -2881 } -2882 -2883 /* -2884 * @return the count of region merge plans executed -2885 */ -2886 public long getMergePlanCount() { -2887return mergePlanCount; -2888 } -2889 -2890 @Override -2891 public boolean registerService(Service instance) { -2892/* -2893 * No stacking of instances is allowed for a single service name -2894 */ -2895Descriptors.ServiceDescriptor serviceDesc = instance.getDescriptorForType(); -2896String serviceName = CoprocessorRpcUtils.getServiceName(serviceDesc); -2897if (coprocessorServiceHandlers.containsKey(serviceName)) { -2898 LOG.error("Coprocessor service "+serviceName+ -2899 " already registered, rejecting request from "+instance -2900 ); -2901 return false; -2902} -2903 -2904 coprocessorServiceHandlers.put(serviceName, instance); -2905if (LOG.isDebugEnabled()) { -2906 LOG.debug("Registered master coprocessor service: service="+serviceName); -2907} -2908return true; -2909 } -2910 -2911 /** -2912 * Utility for constructing an instance of the passed HMaster class. -2913 * @param masterClass -2914 * @return HMaster instance. -2915 */ -2916 public static HMaster constructMaster(Class? extends HMaster masterClass, -2917 final Configuration conf) { -2918try { -2919 Constructor? extends HMaster c = masterClass.getConstructor(Configuration.class); -2920 return c.newInstance(conf); -2921} catch(Exception e) { -2922 Throwable error = e; -2923 if (e instanceof InvocationTargetException -2924 ((InvocationTargetException)e).getTargetException() != null) { -2925error = ((InvocationTargetException)e).getTargetException(); -2926 } -2927 throw new RuntimeException("Failed construction of Master: " + masterClass.toString() + ". " -2928, error); -2929} -2930 } -2931 -2932 /** -2933 * @see org.apache.hadoop.hbase.master.HMasterCommandLine -2934 */ -2935 public static void main(String [] args) { -2936LOG.info("STARTING service " + HMaster.class.getSimpleName()); -2937VersionInfo.logVersion(); -2938new HMasterCommandLine(HMaster.class).doMain(args); -2939 } -2940 -2941 public HFileCleaner getHFileCleaner() { -2942return this.hfileCleaner; -2943 } -2944 -2945 public LogCleaner getLogCleaner() { -2946
[31/51] [partial] hbase-site git commit: Published site at 997747076d8ec0b4346d7cb99c4b0667a7c14905.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/overview-tree.html -- diff --git a/devapidocs/overview-tree.html b/devapidocs/overview-tree.html index 0fea3db..1076efb 100644 --- a/devapidocs/overview-tree.html +++ b/devapidocs/overview-tree.html @@ -2686,6 +2686,7 @@ org.apache.hadoop.hbase.master.assignment.GCMergedRegionsProcedure org.apache.hadoop.hbase.master.assignment.MergeTableRegionsProcedure org.apache.hadoop.hbase.master.procedure.ModifyTableProcedure +org.apache.hadoop.hbase.master.procedure.ReopenTableRegionsProcedure org.apache.hadoop.hbase.master.procedure.RestoreSnapshotProcedure org.apache.hadoop.hbase.master.procedure.TruncateTableProcedure @@ -3629,6 +3630,8 @@ org.apache.hadoop.hbase.client.replication.TableCFs org.apache.hadoop.hbase.client.TableDescriptorBuilder org.apache.hadoop.hbase.client.TableDescriptorBuilder.ModifyableTableDescriptor (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, org.apache.hadoop.hbase.client.TableDescriptor) +org.apache.hadoop.hbase.client.TableDescriptorUtils +org.apache.hadoop.hbase.client.TableDescriptorUtils.TableDescriptorDelta org.apache.hadoop.hbase.rest.model.TableInfoModel (implements org.apache.hadoop.hbase.rest.ProtobufMessageHandler, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) org.apache.hadoop.hbase.mapred.TableInputFormatBase (implements org.apache.hadoop.mapred.InputFormatK,V) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/src-html/org/apache/hadoop/hbase/Version.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html index 655d810..cacb5a3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html @@ -16,11 +16,11 @@ 008@InterfaceAudience.Private 009public class Version { 010 public static final String version = "3.0.0-SNAPSHOT"; -011 public static final String revision = "42be553433775c5985f6c68f8178e51afb0a402e"; +011 public static final String revision = "997747076d8ec0b4346d7cb99c4b0667a7c14905"; 012 public static final String user = "jenkins"; -013 public static final String date = "Tue May 29 14:39:26 UTC 2018"; +013 public static final String date = "Wed May 30 14:39:41 UTC 2018"; 014 public static final String url = "git://jenkins-websites1.apache.org/home/jenkins/jenkins-slave/workspace/hbase_generate_website/hbase"; -015 public static final String srcChecksum = "7bba60c298c587d54706ccb554793a39"; +015 public static final String srcChecksum = "32b36d4fceefb83ecf450c63e8222854"; 016} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/RestoreTablesClient.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/RestoreTablesClient.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/RestoreTablesClient.html index 0cc9507..134fbc3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/RestoreTablesClient.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/RestoreTablesClient.html @@ -85,7 +85,7 @@ 077 /** 078 * Validate target tables. 079 * -080 * @param tTableArray: target tables +080 * @param tTableArray target tables 081 * @param isOverwrite overwrite existing table 082 * @throws IOException exception 083 */ @@ -131,10 +131,10 @@ 123 /** 124 * Restore operation handle each backupImage in array. 125 * -126 * @param images: array BackupImage -127 * @param sTable: table to be restored -128 * @param tTable: table to be restored to -129 * @param truncateIfExists: truncate table +126 * @param images array BackupImage +127 * @param sTable table to be restored +128 * @param tTable table to be restored to +129 * @param truncateIfExists truncate table 130 * @throws IOException exception 131 */ 132 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/src-html/org/apache/hadoop/hbase/client/ClusterConnection.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/ClusterConnection.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/ClusterConnection.html index c1bc8ac..deb8240 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/ClusterConnection.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/ClusterConnection.html @@ -240,7 +240,7 @@ 232 233 /**
[31/51] [partial] hbase-site git commit: Published site at f3d1c021de2264301f68eadb9ef126ff83d7ef53.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/CellComparatorImpl.MetaCellComparator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/CellComparatorImpl.MetaCellComparator.html b/devapidocs/src-html/org/apache/hadoop/hbase/CellComparatorImpl.MetaCellComparator.html index 05d2c3c..6e02e32 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/CellComparatorImpl.MetaCellComparator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/CellComparatorImpl.MetaCellComparator.html @@ -26,441 +26,376 @@ 018 019package org.apache.hadoop.hbase; 020 -021import org.apache.hadoop.hbase.KeyValue.Type; -022import org.apache.hadoop.hbase.util.ByteBufferUtils; -023import org.apache.hadoop.hbase.util.Bytes; -024import org.apache.yetus.audience.InterfaceAudience; -025import org.apache.yetus.audience.InterfaceStability; -026import org.slf4j.Logger; -027import org.slf4j.LoggerFactory; -028 -029import org.apache.hbase.thirdparty.com.google.common.primitives.Longs; -030 -031/** -032 * Compare two HBase cells. Do not use this method comparing code-ROOT-/code or -033 * codehbase:meta/code cells. Cells from these tables need a specialized comparator, one that -034 * takes account of the special formatting of the row where we have commas to delimit table from -035 * regionname, from row. See KeyValue for how it has a special comparator to do hbase:meta cells -036 * and yet another for -ROOT-. -037 * While using this comparator for {{@link #compareRows(Cell, Cell)} et al, the hbase:meta cells -038 * format should be taken into consideration, for which the instance of this comparator -039 * should be used. In all other cases the static APIs in this comparator would be enough -040 */ -041@edu.umd.cs.findbugs.annotations.SuppressWarnings( -042value="UNKNOWN", -043justification="Findbugs doesn't like the way we are negating the result of a compare in below") -044@InterfaceAudience.Private -045@InterfaceStability.Evolving -046public class CellComparatorImpl implements CellComparator { -047 static final Logger LOG = LoggerFactory.getLogger(CellComparatorImpl.class); -048 /** -049 * Comparator for plain key/values; i.e. non-catalog table key/values. Works on Key portion -050 * of KeyValue only. -051 */ -052 public static final CellComparatorImpl COMPARATOR = new CellComparatorImpl(); -053 /** -054 * A {@link CellComparatorImpl} for codehbase:meta/code catalog table -055 * {@link KeyValue}s. -056 */ -057 public static final CellComparatorImpl META_COMPARATOR = new MetaCellComparator(); -058 -059 @Override -060 public int compare(Cell a, Cell b) { -061return compare(a, b, false); -062 } -063 -064 /** -065 * Compare cells. -066 * @param ignoreSequenceid True if we are to compare the key portion only and ignore -067 * the sequenceid. Set to false to compare key and consider sequenceid. -068 * @return 0 if equal, -1 if a lt; b, and +1 if a gt; b. -069 */ -070 @Override -071 public int compare(final Cell a, final Cell b, boolean ignoreSequenceid) { -072int diff = 0; -073if (a instanceof ByteBufferKeyValue b instanceof ByteBufferKeyValue) { -074 diff = compareByteBufferKeyValue((ByteBufferKeyValue)a, (ByteBufferKeyValue)b); -075 if (diff != 0) { -076return diff; -077 } -078} else { -079 diff = compareRows(a, b); -080 if (diff != 0) { -081return diff; -082 } -083 -084 diff = compareWithoutRow(a, b); -085 if (diff != 0) { -086return diff; -087 } -088} -089 -090// Negate following comparisons so later edits show up first mvccVersion: later sorts first -091return ignoreSequenceid? diff: Longs.compare(b.getSequenceId(), a.getSequenceId()); -092 } -093 -094 /** -095 * Specialized comparator for the ByteBufferKeyValue type exclusivesly. -096 * Caches deserialized lengths of rows and families, etc., and reuses them where it can -097 * (ByteBufferKeyValue has been changed to be amenable to our providing pre-made lengths, etc.) -098 */ -099 private static final int compareByteBufferKeyValue(ByteBufferKeyValue left, -100 ByteBufferKeyValue right) { -101// Compare Rows. Cache row length. -102int leftRowLength = left.getRowLength(); -103int rightRowLength = right.getRowLength(); -104int diff = ByteBufferUtils.compareTo( -105left.getRowByteBuffer(), left.getRowPosition(), leftRowLength, -106right.getRowByteBuffer(), right.getRowPosition(), rightRowLength); -107if (diff != 0) { -108 return diff; -109} -110 -111// If the column is not specified, the "minimum" key type appears the -112// latest in the sorted order, regardless of the timestamp. This is used -113// for specifying the last key/value in a given row, because there is no -114//
[31/51] [partial] hbase-site git commit: Published site at cf529f18a9959589fa635f78df4840472526ea2c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html index 5404ea1..1812a55 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html @@ -35,901 +35,908 @@ 027import java.net.InetSocketAddress; 028import java.net.URLDecoder; 029import java.net.URLEncoder; -030import java.util.ArrayList; -031import java.util.Arrays; -032import java.util.List; -033import java.util.Map; -034import java.util.Set; -035import java.util.TreeMap; -036import java.util.TreeSet; -037import java.util.UUID; -038import java.util.function.Function; -039import java.util.stream.Collectors; -040 -041import org.apache.commons.lang3.StringUtils; -042import org.apache.hadoop.conf.Configuration; -043import org.apache.hadoop.fs.FileSystem; -044import org.apache.hadoop.fs.Path; -045import org.apache.hadoop.hbase.Cell; -046import org.apache.hadoop.hbase.CellComparator; -047import org.apache.hadoop.hbase.CellUtil; -048import org.apache.hadoop.hbase.HConstants; -049import org.apache.hadoop.hbase.HRegionLocation; -050import org.apache.hadoop.hbase.HTableDescriptor; -051import org.apache.hadoop.hbase.KeyValue; -052import org.apache.hadoop.hbase.PrivateCellUtil; -053import org.apache.hadoop.hbase.TableName; -054import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -055import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -056import org.apache.hadoop.hbase.client.Connection; -057import org.apache.hadoop.hbase.client.ConnectionFactory; -058import org.apache.hadoop.hbase.client.Put; -059import org.apache.hadoop.hbase.client.RegionLocator; -060import org.apache.hadoop.hbase.client.Table; -061import org.apache.hadoop.hbase.client.TableDescriptor; -062import org.apache.hadoop.hbase.fs.HFileSystem; -063import org.apache.hadoop.hbase.io.ImmutableBytesWritable; -064import org.apache.hadoop.hbase.io.compress.Compression; -065import org.apache.hadoop.hbase.io.compress.Compression.Algorithm; -066import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -067import org.apache.hadoop.hbase.io.hfile.CacheConfig; -068import org.apache.hadoop.hbase.io.hfile.HFile; -069import org.apache.hadoop.hbase.io.hfile.HFileContext; -070import org.apache.hadoop.hbase.io.hfile.HFileContextBuilder; -071import org.apache.hadoop.hbase.io.hfile.HFileWriterImpl; -072import org.apache.hadoop.hbase.regionserver.BloomType; -073import org.apache.hadoop.hbase.regionserver.HStore; -074import org.apache.hadoop.hbase.regionserver.StoreFileWriter; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -077import org.apache.hadoop.hbase.util.FSUtils; -078import org.apache.hadoop.hbase.util.MapReduceExtendedCell; -079import org.apache.hadoop.io.NullWritable; -080import org.apache.hadoop.io.SequenceFile; -081import org.apache.hadoop.io.Text; -082import org.apache.hadoop.mapreduce.Job; -083import org.apache.hadoop.mapreduce.OutputCommitter; -084import org.apache.hadoop.mapreduce.OutputFormat; -085import org.apache.hadoop.mapreduce.RecordWriter; -086import org.apache.hadoop.mapreduce.TaskAttemptContext; -087import org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter; -088import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat; -089import org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner; -090import org.apache.yetus.audience.InterfaceAudience; -091import org.slf4j.Logger; -092import org.slf4j.LoggerFactory; -093 -094import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -095 -096/** -097 * Writes HFiles. Passed Cells must arrive in order. -098 * Writes current time as the sequence id for the file. Sets the major compacted -099 * attribute on created @{link {@link HFile}s. Calling write(null,null) will forcibly roll -100 * all HFiles being written. -101 * p -102 * Using this class as part of a MapReduce job is best done -103 * using {@link #configureIncrementalLoad(Job, TableDescriptor, RegionLocator)}. -104 */ -105@InterfaceAudience.Public -106public class HFileOutputFormat2 -107extends FileOutputFormatImmutableBytesWritable, Cell { -108 private static final Logger LOG = LoggerFactory.getLogger(HFileOutputFormat2.class); -109 static class TableInfo { -110private TableDescriptor tableDesctiptor; -111private RegionLocator regionLocator; -112 -113public TableInfo(TableDescriptor tableDesctiptor, RegionLocator regionLocator) { -114 this.tableDesctiptor = tableDesctiptor; -115 this.regionLocator = regionLocator; -116} -117 -118/**
[31/51] [partial] hbase-site git commit: Published site at 021f66d11d2cbb7308308093e29e69d6e7661ee9.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html index d4390be..1037b84 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactingMemStore.IndexType.html @@ -163,12 +163,12 @@ 155 */ 156 @Override 157 public MemStoreSize size() { -158MemStoreSizing memstoreSizing = new MemStoreSizing(); +158MemStoreSizing memstoreSizing = new NonThreadSafeMemStoreSizing(); 159 memstoreSizing.incMemStoreSize(active.getMemStoreSize()); 160for (Segment item : pipeline.getSegments()) { 161 memstoreSizing.incMemStoreSize(item.getMemStoreSize()); 162} -163return memstoreSizing; +163return memstoreSizing.getMemStoreSize(); 164 } 165 166 /** @@ -224,369 +224,365 @@ 216return new MemStoreSnapshot(snapshotId, this.snapshot); 217 } 218 -219 /** -220 * On flush, how much memory we will clear. -221 * @return size of data that is going to be flushed -222 */ -223 @Override -224 public MemStoreSize getFlushableSize() { -225MemStoreSizing snapshotSizing = getSnapshotSizing(); -226if (snapshotSizing.getDataSize() == 0) { -227 // if snapshot is empty the tail of the pipeline (or everything in the memstore) is flushed -228 if (compositeSnapshot) { -229snapshotSizing = pipeline.getPipelineSizing(); -230 snapshotSizing.incMemStoreSize(active.getMemStoreSize()); -231 } else { -232snapshotSizing = pipeline.getTailSizing(); -233 } -234} -235return snapshotSizing.getDataSize() 0 ? snapshotSizing -236: new MemStoreSize(active.getMemStoreSize()); -237 } -238 -239 @Override -240 protected long keySize() { -241// Need to consider keySize of all segments in pipeline and active -242long k = this.active.keySize(); -243for (Segment segment : this.pipeline.getSegments()) { -244 k += segment.keySize(); -245} -246return k; -247 } -248 -249 @Override -250 protected long heapSize() { -251// Need to consider heapOverhead of all segments in pipeline and active -252long h = this.active.heapSize(); -253for (Segment segment : this.pipeline.getSegments()) { -254 h += segment.heapSize(); -255} -256return h; -257 } -258 -259 @Override -260 public void updateLowestUnflushedSequenceIdInWAL(boolean onlyIfGreater) { -261long minSequenceId = pipeline.getMinSequenceId(); -262if(minSequenceId != Long.MAX_VALUE) { -263 byte[] encodedRegionName = getRegionServices().getRegionInfo().getEncodedNameAsBytes(); -264 byte[] familyName = getFamilyNameInBytes(); -265 WAL WAL = getRegionServices().getWAL(); -266 if (WAL != null) { -267 WAL.updateStore(encodedRegionName, familyName, minSequenceId, onlyIfGreater); -268 } -269} -270 } -271 -272 /** -273 * This message intends to inform the MemStore that next coming updates -274 * are going to be part of the replaying edits from WAL -275 */ -276 @Override -277 public void startReplayingFromWAL() { -278inWalReplay = true; -279 } -280 -281 /** -282 * This message intends to inform the MemStore that the replaying edits from WAL -283 * are done -284 */ -285 @Override -286 public void stopReplayingFromWAL() { -287inWalReplay = false; -288 } -289 -290 // the getSegments() method is used for tests only -291 @VisibleForTesting -292 @Override -293 protected ListSegment getSegments() { -294List? extends Segment pipelineList = pipeline.getSegments(); -295ListSegment list = new ArrayList(pipelineList.size() + 2); -296list.add(this.active); -297list.addAll(pipelineList); -298 list.addAll(this.snapshot.getAllSegments()); -299 -300return list; -301 } -302 -303 // the following three methods allow to manipulate the settings of composite snapshot -304 public void setCompositeSnapshot(boolean useCompositeSnapshot) { -305this.compositeSnapshot = useCompositeSnapshot; -306 } -307 -308 public boolean swapCompactedSegments(VersionedSegmentsList versionedList, ImmutableSegment result, -309 boolean merge) { -310// last true stands for updating the region size -311return pipeline.swap(versionedList, result, !merge, true); -312 } -313 -314 /** -315 * @param requesterVersion The caller must hold the VersionedList of the pipeline -316 * with version taken earlier. This version must be passed as a parameter here. -317 * The flattening happens only if versions match. -318 */ -319 public void
[31/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[31/51] [partial] hbase-site git commit: Published site at 87f5b5f3411d96c31b4cb61b9a57ced22be91d1f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html index e1bc325..63e7421 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.FileLockCallable.html @@ -66,5125 +66,5224 @@ 058import java.util.concurrent.TimeoutException; 059import java.util.concurrent.atomic.AtomicBoolean; 060import java.util.concurrent.atomic.AtomicInteger; -061import org.apache.commons.io.IOUtils; -062import org.apache.commons.lang3.RandomStringUtils; -063import org.apache.commons.lang3.StringUtils; -064import org.apache.hadoop.conf.Configuration; -065import org.apache.hadoop.conf.Configured; -066import org.apache.hadoop.fs.FSDataOutputStream; -067import org.apache.hadoop.fs.FileStatus; -068import org.apache.hadoop.fs.FileSystem; -069import org.apache.hadoop.fs.Path; -070import org.apache.hadoop.fs.permission.FsAction; -071import org.apache.hadoop.fs.permission.FsPermission; -072import org.apache.hadoop.hbase.Abortable; -073import org.apache.hadoop.hbase.Cell; -074import org.apache.hadoop.hbase.CellUtil; -075import org.apache.hadoop.hbase.ClusterMetrics; -076import org.apache.hadoop.hbase.ClusterMetrics.Option; -077import org.apache.hadoop.hbase.HBaseConfiguration; -078import org.apache.hadoop.hbase.HBaseInterfaceAudience; -079import org.apache.hadoop.hbase.HConstants; -080import org.apache.hadoop.hbase.HRegionInfo; -081import org.apache.hadoop.hbase.HRegionLocation; -082import org.apache.hadoop.hbase.KeyValue; -083import org.apache.hadoop.hbase.MasterNotRunningException; -084import org.apache.hadoop.hbase.MetaTableAccessor; -085import org.apache.hadoop.hbase.RegionLocations; -086import org.apache.hadoop.hbase.ServerName; -087import org.apache.hadoop.hbase.TableName; -088import org.apache.hadoop.hbase.ZooKeeperConnectionException; -089import org.apache.hadoop.hbase.client.Admin; -090import org.apache.hadoop.hbase.client.ClusterConnection; -091import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -092import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -093import org.apache.hadoop.hbase.client.Connection; -094import org.apache.hadoop.hbase.client.ConnectionFactory; -095import org.apache.hadoop.hbase.client.Delete; -096import org.apache.hadoop.hbase.client.Get; -097import org.apache.hadoop.hbase.client.Put; -098import org.apache.hadoop.hbase.client.RegionInfo; -099import org.apache.hadoop.hbase.client.RegionInfoBuilder; -100import org.apache.hadoop.hbase.client.RegionReplicaUtil; -101import org.apache.hadoop.hbase.client.Result; -102import org.apache.hadoop.hbase.client.RowMutations; -103import org.apache.hadoop.hbase.client.Table; -104import org.apache.hadoop.hbase.client.TableDescriptor; -105import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -106import org.apache.hadoop.hbase.client.TableState; -107import org.apache.hadoop.hbase.io.FileLink; -108import org.apache.hadoop.hbase.io.HFileLink; -109import org.apache.hadoop.hbase.io.hfile.CacheConfig; -110import org.apache.hadoop.hbase.io.hfile.HFile; -111import org.apache.hadoop.hbase.log.HBaseMarkers; -112import org.apache.hadoop.hbase.master.MasterFileSystem; -113import org.apache.hadoop.hbase.master.RegionState; -114import org.apache.hadoop.hbase.regionserver.HRegion; -115import org.apache.hadoop.hbase.regionserver.HRegionFileSystem; -116import org.apache.hadoop.hbase.regionserver.StoreFileInfo; -117import org.apache.hadoop.hbase.replication.ReplicationException; -118import org.apache.hadoop.hbase.security.AccessDeniedException; -119import org.apache.hadoop.hbase.security.UserProvider; -120import org.apache.hadoop.hbase.util.Bytes.ByteArrayComparator; -121import org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE; -122import org.apache.hadoop.hbase.util.hbck.HFileCorruptionChecker; -123import org.apache.hadoop.hbase.util.hbck.ReplicationChecker; -124import org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandler; -125import org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandlerImpl; -126import org.apache.hadoop.hbase.wal.WAL; -127import org.apache.hadoop.hbase.wal.WALFactory; -128import org.apache.hadoop.hbase.wal.WALSplitter; -129import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; -130import org.apache.hadoop.hbase.zookeeper.ZKUtil; -131import org.apache.hadoop.hbase.zookeeper.ZKWatcher; -132import org.apache.hadoop.hbase.zookeeper.ZNodePaths; -133import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException; -134import org.apache.hadoop.ipc.RemoteException; -135import
[31/51] [partial] hbase-site git commit: Published site at 2912c953551bedbfbf30c32c156ed7bb187d54c3.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html index b862507..b6e7636 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html @@ -4235,7 +4235,22 @@ 4227 } 4228}); 4229 } -4230} +4230 +4231 @Override +4232 public void cloneTableSchema(final TableName tableName, final TableName newTableName, +4233 final boolean preserveSplits) throws IOException { +4234checkTableExists(tableName); +4235if (tableExists(newTableName)) { +4236 throw new TableExistsException(newTableName); +4237} +4238TableDescriptor htd = TableDescriptorBuilder.copy(newTableName, getTableDescriptor(tableName)); +4239if (preserveSplits) { +4240 createTable(htd, getTableSplits(tableName)); +4241} else { +4242 createTable(htd); +4243} +4244 } +4245} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html index b862507..b6e7636 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html @@ -4235,7 +4235,22 @@ 4227 } 4228}); 4229 } -4230} +4230 +4231 @Override +4232 public void cloneTableSchema(final TableName tableName, final TableName newTableName, +4233 final boolean preserveSplits) throws IOException { +4234checkTableExists(tableName); +4235if (tableExists(newTableName)) { +4236 throw new TableExistsException(newTableName); +4237} +4238TableDescriptor htd = TableDescriptorBuilder.copy(newTableName, getTableDescriptor(tableName)); +4239if (preserveSplits) { +4240 createTable(htd, getTableSplits(tableName)); +4241} else { +4242 createTable(htd); +4243} +4244 } +4245} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html index b862507..b6e7636 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html @@ -4235,7 +4235,22 @@ 4227 } 4228}); 4229 } -4230} +4230 +4231 @Override +4232 public void cloneTableSchema(final TableName tableName, final TableName newTableName, +4233 final boolean preserveSplits) throws IOException { +4234checkTableExists(tableName); +4235if (tableExists(newTableName)) { +4236 throw new TableExistsException(newTableName); +4237} +4238TableDescriptor htd = TableDescriptorBuilder.copy(newTableName, getTableDescriptor(tableName)); +4239if (preserveSplits) { +4240 createTable(htd, getTableSplits(tableName)); +4241} else { +4242 createTable(htd); +4243} +4244 } +4245} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html index b862507..b6e7636 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html @@ -4235,7 +4235,22 @@ 4227 } 4228}); 4229 } -4230} +4230 +4231 @Override +4232 public void cloneTableSchema(final TableName tableName, final TableName newTableName, +4233 final boolean preserveSplits) throws IOException { +4234checkTableExists(tableName); +4235if (tableExists(newTableName)) { +4236 throw new TableExistsException(newTableName); +4237} +4238TableDescriptor htd =
[31/51] [partial] hbase-site git commit: Published site at 2a2258656b2fcd92b967131b6c1f037363553bc4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index 85bb6c9..750a1f2 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -704,20 +704,20 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) +org.apache.hadoop.hbase.regionserver.ScanType +org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result +org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType +org.apache.hadoop.hbase.regionserver.BloomType org.apache.hadoop.hbase.regionserver.FlushType -org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection -org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType -org.apache.hadoop.hbase.regionserver.ScannerContext.NextState +org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type org.apache.hadoop.hbase.regionserver.Region.Operation +org.apache.hadoop.hbase.regionserver.ScannerContext.NextState org.apache.hadoop.hbase.regionserver.MemStoreCompactionStrategy.Action -org.apache.hadoop.hbase.regionserver.ScanType -org.apache.hadoop.hbase.regionserver.BloomType -org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status -org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope -org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html index 858ccf6..3bd22b5 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html @@ -130,9 +130,9 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult -org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode +org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput +org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/replication/ReplicationQueueStorage.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/replication/ReplicationQueueStorage.html b/devapidocs/org/apache/hadoop/hbase/replication/ReplicationQueueStorage.html index d7a6f7f..6b44a8e 100644 --- a/devapidocs/org/apache/hadoop/hbase/replication/ReplicationQueueStorage.html +++ b/devapidocs/org/apache/hadoop/hbase/replication/ReplicationQueueStorage.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6}; +var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"]}; var altColor =
[31/51] [partial] hbase-site git commit: Published site at e468b4022f76688851b3e0c34722f01a56bd624f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/apidocs/src-html/org/apache/hadoop/hbase/client/Get.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Get.html b/apidocs/src-html/org/apache/hadoop/hbase/client/Get.html index b3689ed..bfec785 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Get.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Get.html @@ -64,7 +64,7 @@ 056 * execute {@link #setTimeRange(long, long) setTimeRange}. 057 * p 058 * To only retrieve columns with a specific timestamp, execute -059 * {@link #setTimeStamp(long) setTimestamp}. +059 * {@link #setTimestamp(long) setTimestamp}. 060 * p 061 * To limit the number of versions of each column to be returned, execute 062 * {@link #setMaxVersions(int) setMaxVersions}. @@ -80,7 +80,7 @@ 072 private boolean cacheBlocks = true; 073 private int storeLimit = -1; 074 private int storeOffset = 0; -075 private TimeRange tr = new TimeRange(); +075 private TimeRange tr = TimeRange.allTime(); 076 private boolean checkExistenceOnly = false; 077 private boolean closestRowBefore = false; 078 private Mapbyte [], NavigableSetbyte [] familyMap = new TreeMap(Bytes.BYTES_COMPARATOR); @@ -239,350 +239,362 @@ 231 * Get versions of columns with the specified timestamp. 232 * @param timestamp version timestamp 233 * @return this for invocation chaining -234 */ -235 public Get setTimeStamp(long timestamp) -236 throws IOException { -237try { -238 tr = new TimeRange(timestamp, timestamp+1); -239} catch(Exception e) { -240 // This should never happen, unless integer overflow or something extremely wrong... -241 LOG.error("TimeRange failed, likely caused by integer overflow. ", e); -242 throw e; -243} -244return this; -245 } -246 -247 @Override public Get setColumnFamilyTimeRange(byte[] cf, long minStamp, long maxStamp) { -248return (Get) super.setColumnFamilyTimeRange(cf, minStamp, maxStamp); -249 } -250 -251 /** -252 * Get all available versions. -253 * @return this for invocation chaining -254 * @deprecated It is easy to misunderstand with column family's max versions, so use -255 * {@link #readAllVersions()} instead. -256 */ -257 @Deprecated -258 public Get setMaxVersions() { -259return readAllVersions(); -260 } -261 -262 /** -263 * Get up to the specified number of versions of each column. -264 * @param maxVersions maximum versions for each column -265 * @throws IOException if invalid number of versions -266 * @return this for invocation chaining -267 * @deprecated It is easy to misunderstand with column family's max versions, so use -268 * {@link #readVersions(int)} instead. -269 */ -270 @Deprecated -271 public Get setMaxVersions(int maxVersions) throws IOException { -272return readVersions(maxVersions); -273 } -274 -275 /** -276 * Get all available versions. -277 * @return this for invocation chaining -278 */ -279 public Get readAllVersions() { -280this.maxVersions = Integer.MAX_VALUE; -281return this; -282 } -283 -284 /** -285 * Get up to the specified number of versions of each column. -286 * @param versions specified number of versions for each column -287 * @throws IOException if invalid number of versions -288 * @return this for invocation chaining -289 */ -290 public Get readVersions(int versions) throws IOException { -291if (versions = 0) { -292 throw new IOException("versions must be positive"); -293} -294this.maxVersions = versions; -295return this; -296 } -297 -298 @Override -299 public Get setLoadColumnFamiliesOnDemand(boolean value) { -300return (Get) super.setLoadColumnFamiliesOnDemand(value); -301 } -302 -303 /** -304 * Set the maximum number of values to return per row per Column Family -305 * @param limit the maximum number of values returned / row / CF -306 * @return this for invocation chaining -307 */ -308 public Get setMaxResultsPerColumnFamily(int limit) { -309this.storeLimit = limit; -310return this; -311 } -312 -313 /** -314 * Set offset for the row per Column Family. This offset is only within a particular row/CF -315 * combination. It gets reset back to zero when we move to the next row or CF. -316 * @param offset is the number of kvs that will be skipped. -317 * @return this for invocation chaining -318 */ -319 public Get setRowOffsetPerColumnFamily(int offset) { -320this.storeOffset = offset; -321return this; -322 } -323 -324 @Override -325 public Get setFilter(Filter filter) { -326super.setFilter(filter); -327return this; -328 } -329 -330 /* Accessors */ -331 -332 /** -333 * Set whether blocks should be cached for this Get. -334 * p -335 * This is true by default. When true, default settings of the table and
[31/51] [partial] hbase-site git commit: Published site at 64061f896fe21512504e3886a400759e88b519da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html b/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html index 313a472..1db7ad2 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html @@ -816,30 +816,25 @@ void -FavoredStochasticBalancer.initialize() +BaseLoadBalancer.initialize() void -BaseLoadBalancer.initialize() +FavoredStochasticBalancer.initialize() ServerName -FavoredStochasticBalancer.randomAssignment(RegionInforegionInfo, -https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) - - -ServerName BaseLoadBalancer.randomAssignment(RegionInforegionInfo, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) Used to assign a single region to a random server. - -https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -FavoredStochasticBalancer.retainAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapRegionInfo,ServerNameregions, + +ServerName +FavoredStochasticBalancer.randomAssignment(RegionInforegionInfo, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) - + https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo BaseLoadBalancer.retainAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapRegionInfo,ServerNameregions, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) @@ -848,12 +843,12 @@ available/online servers available for assignment. - + https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -FavoredStochasticBalancer.roundRobinAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegions, -https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) +FavoredStochasticBalancer.retainAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapRegionInfo,ServerNameregions, +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) - + https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo BaseLoadBalancer.roundRobinAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegions, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) @@ -861,6 +856,11 @@ simple round-robin assignment. + +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +FavoredStochasticBalancer.roundRobinAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegions, +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) + private
[31/51] [partial] hbase-site git commit: Published site at 4cb40e6d846ce1f28ffb40d388c9efb753197813.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/class-use/ClusterMetrics.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ClusterMetrics.html b/devapidocs/org/apache/hadoop/hbase/class-use/ClusterMetrics.html index 1621237..11d7754 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/ClusterMetrics.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/ClusterMetrics.html @@ -242,27 +242,27 @@ https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics -AsyncHBaseAdmin.getClusterMetrics() +AsyncAdmin.getClusterMetrics() https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics -AsyncAdmin.getClusterMetrics() +RawAsyncHBaseAdmin.getClusterMetrics() https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics -RawAsyncHBaseAdmin.getClusterMetrics() +AsyncHBaseAdmin.getClusterMetrics() https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics -AsyncHBaseAdmin.getClusterMetrics(https://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true; title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions) +AsyncAdmin.getClusterMetrics(https://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true; title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions) https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics -AsyncAdmin.getClusterMetrics(https://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true; title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions) +RawAsyncHBaseAdmin.getClusterMetrics(https://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true; title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions) https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics -RawAsyncHBaseAdmin.getClusterMetrics(https://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true; title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions) +AsyncHBaseAdmin.getClusterMetrics(https://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true; title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions) @@ -408,11 +408,11 @@ void -BaseLoadBalancer.setClusterMetrics(ClusterMetricsst) +RegionLocationFinder.setClusterMetrics(ClusterMetricsstatus) void -RegionLocationFinder.setClusterMetrics(ClusterMetricsstatus) +BaseLoadBalancer.setClusterMetrics(ClusterMetricsst) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html b/devapidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html index 804b5df..a8beae7 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html @@ -186,94 +186,94 @@ the order they are declared. boolean -Table.checkAndDelete(byte[]row, +HTable.checkAndDelete(byte[]row, byte[]family, byte[]qualifier, CompareOperatorop, byte[]value, Deletedelete) -Deprecated. -Since 2.0.0. Will be removed in 3.0.0. Use Table.checkAndMutate(byte[], byte[]) - +Deprecated. boolean -HTable.checkAndDelete(byte[]row, +Table.checkAndDelete(byte[]row, byte[]family, byte[]qualifier, CompareOperatorop, byte[]value, Deletedelete) -Deprecated. +Deprecated. +Since 2.0.0. Will be removed in 3.0.0. Use Table.checkAndMutate(byte[], byte[]) + boolean -Table.checkAndMutate(byte[]row, +HTable.checkAndMutate(byte[]row, byte[]family, byte[]qualifier, CompareOperatorop, byte[]value, - RowMutationsmutation) -Deprecated. -Since 2.0.0. Will be removed in 3.0.0. Use Table.checkAndMutate(byte[], byte[]) - +
[31/51] [partial] hbase-site git commit: Published site at 8ab7b20f48951d77945181024f5e15842bc253c4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html index ecf500c..0cd5a4e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html @@ -238,8355 +238,8368 @@ 230 public static final String HBASE_MAX_CELL_SIZE_KEY = "hbase.server.keyvalue.maxsize"; 231 public static final int DEFAULT_MAX_CELL_SIZE = 10485760; 232 -233 public static final String HBASE_REGIONSERVER_MINIBATCH_SIZE = -234 "hbase.regionserver.minibatch.size"; -235 public static final int DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2; -236 -237 /** -238 * This is the global default value for durability. All tables/mutations not -239 * defining a durability or using USE_DEFAULT will default to this value. -240 */ -241 private static final Durability DEFAULT_DURABILITY = Durability.SYNC_WAL; +233 /** +234 * This is the global default value for durability. All tables/mutations not +235 * defining a durability or using USE_DEFAULT will default to this value. +236 */ +237 private static final Durability DEFAULT_DURABILITY = Durability.SYNC_WAL; +238 +239 public static final String HBASE_REGIONSERVER_MINIBATCH_SIZE = +240 "hbase.regionserver.minibatch.size"; +241 public static final int DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2; 242 -243 final AtomicBoolean closed = new AtomicBoolean(false); -244 -245 /* Closing can take some time; use the closing flag if there is stuff we don't -246 * want to do while in closing state; e.g. like offer this region up to the -247 * master as a region to close if the carrying regionserver is overloaded. -248 * Once set, it is never cleared. -249 */ -250 final AtomicBoolean closing = new AtomicBoolean(false); -251 -252 /** -253 * The max sequence id of flushed data on this region. There is no edit in memory that is -254 * less that this sequence id. -255 */ -256 private volatile long maxFlushedSeqId = HConstants.NO_SEQNUM; -257 -258 /** -259 * Record the sequence id of last flush operation. Can be in advance of -260 * {@link #maxFlushedSeqId} when flushing a single column family. In this case, -261 * {@link #maxFlushedSeqId} will be older than the oldest edit in memory. -262 */ -263 private volatile long lastFlushOpSeqId = HConstants.NO_SEQNUM; -264 -265 /** -266 * The sequence id of the last replayed open region event from the primary region. This is used -267 * to skip entries before this due to the possibility of replay edits coming out of order from -268 * replication. -269 */ -270 protected volatile long lastReplayedOpenRegionSeqId = -1L; -271 protected volatile long lastReplayedCompactionSeqId = -1L; -272 -273 // -274 // Members -275 // -276 -277 // map from a locked row to the context for that lock including: -278 // - CountDownLatch for threads waiting on that row -279 // - the thread that owns the lock (allow reentrancy) -280 // - reference count of (reentrant) locks held by the thread -281 // - the row itself -282 private final ConcurrentHashMapHashedBytes, RowLockContext lockedRows = -283 new ConcurrentHashMap(); -284 -285 protected final Mapbyte[], HStore stores = -286 new ConcurrentSkipListMap(Bytes.BYTES_RAWCOMPARATOR); +243 public static final String WAL_HSYNC_CONF_KEY = "hbase.wal.hsync"; +244 public static final boolean DEFAULT_WAL_HSYNC = false; +245 +246 final AtomicBoolean closed = new AtomicBoolean(false); +247 +248 /* Closing can take some time; use the closing flag if there is stuff we don't +249 * want to do while in closing state; e.g. like offer this region up to the +250 * master as a region to close if the carrying regionserver is overloaded. +251 * Once set, it is never cleared. +252 */ +253 final AtomicBoolean closing = new AtomicBoolean(false); +254 +255 /** +256 * The max sequence id of flushed data on this region. There is no edit in memory that is +257 * less that this sequence id. +258 */ +259 private volatile long maxFlushedSeqId = HConstants.NO_SEQNUM; +260 +261 /** +262 * Record the sequence id of last flush operation. Can be in advance of +263 * {@link #maxFlushedSeqId} when flushing a single column family. In this case, +264 * {@link #maxFlushedSeqId} will be older than the oldest edit in memory. +265 */ +266 private volatile long lastFlushOpSeqId =
[31/51] [partial] hbase-site git commit: Published site at 00095a2ef9442e3fd86c04876c9d91f2f8b23ad8.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.FileConverter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.FileConverter.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.FileConverter.html index f47d627..c3d225c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.FileConverter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.FileConverter.html @@ -117,219 +117,219 @@ 109 */ 110 public static boolean archiveRegion(FileSystem fs, Path rootdir, Path tableDir, Path regionDir) 111 throws IOException { -112if (LOG.isDebugEnabled()) { -113 LOG.debug("ARCHIVING " + regionDir.toString()); -114} -115 -116// otherwise, we archive the files -117// make sure we can archive -118if (tableDir == null || regionDir == null) { -119 LOG.error("No archive directory could be found because tabledir (" + tableDir -120 + ") or regiondir (" + regionDir + "was null. Deleting files instead."); -121 deleteRegionWithoutArchiving(fs, regionDir); -122 // we should have archived, but failed to. Doesn't matter if we deleted -123 // the archived files correctly or not. -124 return false; -125} -126 -127// make sure the regiondir lives under the tabledir -128 Preconditions.checkArgument(regionDir.toString().startsWith(tableDir.toString())); -129Path regionArchiveDir = HFileArchiveUtil.getRegionArchiveDir(rootdir, -130FSUtils.getTableName(tableDir), -131regionDir.getName()); -132 -133FileStatusConverter getAsFile = new FileStatusConverter(fs); -134// otherwise, we attempt to archive the store files -135 -136// build collection of just the store directories to archive -137CollectionFile toArchive = new ArrayList(); -138final PathFilter dirFilter = new FSUtils.DirFilter(fs); -139PathFilter nonHidden = new PathFilter() { -140 @Override -141 public boolean accept(Path file) { -142return dirFilter.accept(file) !file.getName().toString().startsWith("."); -143 } -144}; -145FileStatus[] storeDirs = FSUtils.listStatus(fs, regionDir, nonHidden); -146// if there no files, we can just delete the directory and return; -147if (storeDirs == null) { -148 LOG.debug("Region directory " + regionDir + " empty."); -149 return deleteRegionWithoutArchiving(fs, regionDir); -150} -151 -152// convert the files in the region to a File -153 toArchive.addAll(Lists.transform(Arrays.asList(storeDirs), getAsFile)); -154LOG.debug("Archiving " + toArchive); -155ListFile failedArchive = resolveAndArchive(fs, regionArchiveDir, toArchive, -156 EnvironmentEdgeManager.currentTime()); -157if (!failedArchive.isEmpty()) { -158 throw new FailedArchiveException("Failed to archive/delete all the files for region:" -159 + regionDir.getName() + " into " + regionArchiveDir -160 + ". Something is probably awry on the filesystem.", -161 Collections2.transform(failedArchive, FUNC_FILE_TO_PATH)); -162} -163// if that was successful, then we delete the region -164return deleteRegionWithoutArchiving(fs, regionDir); -165 } -166 -167 /** -168 * Remove from the specified region the store files of the specified column family, -169 * either by archiving them or outright deletion -170 * @param fs the filesystem where the store files live -171 * @param conf {@link Configuration} to examine to determine the archive directory -172 * @param parent Parent region hosting the store files -173 * @param tableDir {@link Path} to where the table is being stored (for building the archive path) -174 * @param family the family hosting the store files -175 * @throws IOException if the files could not be correctly disposed. -176 */ -177 public static void archiveFamily(FileSystem fs, Configuration conf, -178 RegionInfo parent, Path tableDir, byte[] family) throws IOException { -179Path familyDir = new Path(tableDir, new Path(parent.getEncodedName(), Bytes.toString(family))); -180archiveFamilyByFamilyDir(fs, conf, parent, familyDir, family); -181 } -182 -183 /** -184 * Removes from the specified region the store files of the specified column family, -185 * either by archiving them or outright deletion -186 * @param fs the filesystem where the store files live -187 * @param conf {@link Configuration} to examine to determine the archive directory -188 * @param parent Parent region hosting the store files -189 * @param familyDir {@link Path} to where the family is being stored -190 * @param family the family hosting the store files -191 * @throws IOException if the files could not be correctly disposed.
[31/51] [partial] hbase-site git commit: Published site at 22f4def942f8a3367d0ca6598317e9b9a7d0cfcd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html index 6c32c5c..3c17832 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html +++ b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":9,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":42,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":9,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":42,"i52":42,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":9,"i64":10,"i65":10,"i66":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":42,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":9,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":42,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":42,"i51":42,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":9,"i63":10,"i64":10,"i65":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public static class TableDescriptorBuilder.ModifyableTableDescriptor +public static class TableDescriptorBuilder.ModifyableTableDescriptor extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements TableDescriptor, https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableTableDescriptorBuilder.ModifyableTableDescriptor TODO: make this private after removing the HTableDescriptor @@ -217,395 +217,391 @@ implements Method and Description -TableDescriptorBuilder.ModifyableTableDescriptor -addColumnFamily(ColumnFamilyDescriptorfamily) -Adds a column family. - - - -TableDescriptorBuilder.ModifyableTableDescriptor -addCoprocessor(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringclassName) -Add a table coprocessor to this table. - - - -TableDescriptorBuilder.ModifyableTableDescriptor -addCoprocessor(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringclassName, - org.apache.hadoop.fs.PathjarFilePath, - intpriority, - https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringkvs) -Add a table coprocessor to this table. - - - -private TableDescriptorBuilder.ModifyableTableDescriptor -addCoprocessorToMap(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringspecStr) -Add coprocessor to values Map - - - -TableDescriptorBuilder.ModifyableTableDescriptor -addCoprocessorWithSpec(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringspecStr) -Add a table coprocessor to this table. - - - private void checkHasCoprocessor(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringclassName) - + int compareTo(TableDescriptorBuilder.ModifyableTableDescriptorother) Compares the descriptor with another descriptor which is passed as a parameter. - + boolean equals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectobj) Compare the
[31/51] [partial] hbase-site git commit: Published site at 31da4d0bce69b3a47066a5df675756087ce4dc60.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html b/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html index 5f0d8e8..a7531a8 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html +++ b/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html @@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class MasterRpcServices +public class MasterRpcServices extends RSRpcServices implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterService.BlockingInterface, org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStatusService.BlockingInterface, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockService.BlockingInterface Implements the master RPC services. @@ -174,7 +174,7 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master Fields inherited from classorg.apache.hadoop.hbase.regionserver.RSRpcServices -REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS, REGIONSERVER_ADMIN_SERVICE_CONFIG, REGIONSERVER_CLIENT_SERVICE_CONFIG, RESERVOIR_ENABLED_KEY +accessChecker, REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS, REGIONSERVER_ADMIN_SERVICE_CONFIG, REGIONSERVER_CLIENT_SERVICE_CONFIG, RESERVOIR_ENABLED_KEY @@ -793,7 +793,7 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -802,7 +802,7 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master master -private finalHMaster master +private finalHMaster master @@ -819,7 +819,7 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master MasterRpcServices -publicMasterRpcServices(HMasterm) +publicMasterRpcServices(HMasterm) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -841,7 +841,7 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master createConfigurationSubset -privateorg.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse.BuildercreateConfigurationSubset() +privateorg.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse.BuildercreateConfigurationSubset() Returns: Subset of configuration to pass initializing regionservers: e.g. @@ -855,7 +855,7 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master addConfig -privateorg.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse.BuilderaddConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse.Builderresp, +privateorg.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse.BuilderaddConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupResponse.Builderresp, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringkey) @@ -865,7 +865,7 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master createRpcServer -protectedRpcServerInterfacecreateRpcServer(Serverserver, +protectedRpcServerInterfacecreateRpcServer(Serverserver, org.apache.hadoop.conf.Configurationconf, RpcSchedulerFactoryrpcSchedulerFactory, https://docs.oracle.com/javase/8/docs/api/java/net/InetSocketAddress.html?is-external=true; title="class or interface in java.net">InetSocketAddressbindAddress, @@ -885,7 +885,7 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master createPriority -protectedPriorityFunctioncreatePriority() +protectedPriorityFunctioncreatePriority() Overrides: createPriorityin classRSRpcServices @@ -898,7 +898,7 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master rpcPreCheck -privatevoidrpcPreCheck(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringrequestName) +privatevoidrpcPreCheck(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in
[31/51] [partial] hbase-site git commit: Published site at 6b77786dfc46d25ac5bb5f1c8a4a9eb47b52a604.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html b/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html index 44ca7f5..f2a249c 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html @@ -239,15 +239,15 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private ServerName -HRegionLocation.serverName +ServerMetricsBuilder.serverName private ServerName -ServerMetricsBuilder.serverName +ServerMetricsBuilder.ServerMetricsImpl.serverName private ServerName -ServerMetricsBuilder.ServerMetricsImpl.serverName +HRegionLocation.serverName @@ -306,9 +306,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. ServerName -ClusterMetrics.getMasterName() -Returns detailed information about the current master ServerName. - +ClusterMetricsBuilder.ClusterMetricsImpl.getMasterName() ServerName @@ -318,11 +316,15 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. ServerName -ClusterMetricsBuilder.ClusterMetricsImpl.getMasterName() +ClusterMetrics.getMasterName() +Returns detailed information about the current master ServerName. + ServerName -HRegionLocation.getServerName() +ServerLoad.getServerName() +Deprecated. + ServerName @@ -330,13 +332,11 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. ServerName -ServerLoad.getServerName() -Deprecated. - +ServerMetricsBuilder.ServerMetricsImpl.getServerName() ServerName -ServerMetricsBuilder.ServerMetricsImpl.getServerName() +HRegionLocation.getServerName() ServerName @@ -405,7 +405,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName -ClusterMetrics.getBackupMasterNames() +ClusterMetricsBuilder.ClusterMetricsImpl.getBackupMasterNames() https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName @@ -415,7 +415,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName -ClusterMetricsBuilder.ClusterMetricsImpl.getBackupMasterNames() +ClusterMetrics.getBackupMasterNames() https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName @@ -428,7 +428,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName -ClusterMetrics.getDeadServerNames() +ClusterMetricsBuilder.ClusterMetricsImpl.getDeadServerNames() https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName @@ -438,7 +438,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName -ClusterMetricsBuilder.ClusterMetricsImpl.getDeadServerNames() +ClusterMetrics.getDeadServerNames() private https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,ServerLoad @@ -448,7 +448,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,ServerMetrics -ClusterMetrics.getLiveServerMetrics() +ClusterMetricsBuilder.ClusterMetricsImpl.getLiveServerMetrics() https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,ServerMetrics @@ -458,7 +458,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,ServerMetrics -ClusterMetricsBuilder.ClusterMetricsImpl.getLiveServerMetrics() +ClusterMetrics.getLiveServerMetrics() static PairRegionInfo,ServerName @@ -857,31 +857,31 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private ServerName -AsyncRequestFutureImpl.SingleServerRequestRunnable.server +FastFailInterceptorContext.server private ServerName -FastFailInterceptorContext.server
[31/51] [partial] hbase-site git commit: Published site at 1384da71375427b522b09f06862bb5d629cef52f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html b/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html index bdc486d..cdc563c 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/HBaseIOException.html @@ -796,25 +796,30 @@ void -BaseLoadBalancer.initialize() +FavoredStochasticBalancer.initialize() void -FavoredStochasticBalancer.initialize() +BaseLoadBalancer.initialize() ServerName +FavoredStochasticBalancer.randomAssignment(RegionInforegionInfo, +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) + + +ServerName BaseLoadBalancer.randomAssignment(RegionInforegionInfo, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) Used to assign a single region to a random server. - -ServerName -FavoredStochasticBalancer.randomAssignment(RegionInforegionInfo, + +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +FavoredStochasticBalancer.retainAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapRegionInfo,ServerNameregions, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) - + https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo BaseLoadBalancer.retainAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapRegionInfo,ServerNameregions, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) @@ -823,12 +828,12 @@ available/online servers available for assignment. - + https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -FavoredStochasticBalancer.retainAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapRegionInfo,ServerNameregions, -https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) +FavoredStochasticBalancer.roundRobinAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegions, +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) - + https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo BaseLoadBalancer.roundRobinAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegions, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) @@ -836,11 +841,6 @@ simple round-robin assignment. - -https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -FavoredStochasticBalancer.roundRobinAssignment(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegions, -https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) - private
[31/51] [partial] hbase-site git commit: Published site at b7b86839250bf9b295ebc1948826f43a88736d6c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/org/apache/hadoop/hbase/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/package-tree.html b/devapidocs/org/apache/hadoop/hbase/package-tree.html index 6791ccc..73d0c00 100644 --- a/devapidocs/org/apache/hadoop/hbase/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/package-tree.html @@ -222,6 +222,7 @@ org.apache.hadoop.hbase.MetaTableAccessor.TableVisitorBase +org.apache.hadoop.hbase.MetaTableAccessor.ReplicationBarrierResult org.apache.hadoop.hbase.NamespaceDescriptor org.apache.hadoop.hbase.NamespaceDescriptor.Builder org.apache.hadoop.hbase.PrivateCellUtil @@ -440,20 +441,20 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.CompareOperator -org.apache.hadoop.hbase.KeyValue.Type -org.apache.hadoop.hbase.ProcedureState -org.apache.hadoop.hbase.Cell.Type -org.apache.hadoop.hbase.KeepDeletedCells org.apache.hadoop.hbase.HealthChecker.HealthCheckerExitStatus org.apache.hadoop.hbase.MemoryCompactionPolicy +org.apache.hadoop.hbase.CompareOperator +org.apache.hadoop.hbase.ClusterMetrics.Option org.apache.hadoop.hbase.Size.Unit org.apache.hadoop.hbase.CompatibilitySingletonFactory.SingletonStorage +org.apache.hadoop.hbase.ProcedureState +org.apache.hadoop.hbase.HConstants.OperationStatusCode org.apache.hadoop.hbase.Coprocessor.State -org.apache.hadoop.hbase.ClusterMetrics.Option org.apache.hadoop.hbase.CellBuilderType -org.apache.hadoop.hbase.HConstants.OperationStatusCode +org.apache.hadoop.hbase.KeepDeletedCells org.apache.hadoop.hbase.MetaTableAccessor.QueryType +org.apache.hadoop.hbase.Cell.Type +org.apache.hadoop.hbase.KeyValue.Type http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/org/apache/hadoop/hbase/package-use.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/package-use.html b/devapidocs/org/apache/hadoop/hbase/package-use.html index 266e8a7..fae3549 100644 --- a/devapidocs/org/apache/hadoop/hbase/package-use.html +++ b/devapidocs/org/apache/hadoop/hbase/package-use.html @@ -695,72 +695,75 @@ service. MetaTableAccessor.QueryType +MetaTableAccessor.ReplicationBarrierResult + + MetaTableAccessor.Visitor Implementations 'visit' a catalog table row. - + NamespaceDescriptor Namespace POJO class. - + NamespaceDescriptor.Builder - + PrivateCellUtil.EmptyByteBufferExtendedCell These cells are used in reseeks/seeks to improve the read performance. - + PrivateCellUtil.EmptyCell These cells are used in reseeks/seeks to improve the read performance. - + PrivateCellUtil.FirstOnRowByteBufferExtendedCell - + PrivateCellUtil.FirstOnRowCell - + PrivateCellUtil.FirstOnRowColByteBufferExtendedCell - + PrivateCellUtil.FirstOnRowColCell - + PrivateCellUtil.LastOnRowByteBufferExtendedCell - + PrivateCellUtil.LastOnRowCell - + PrivateCellUtil.TagRewriteByteBufferExtendedCell - + PrivateCellUtil.TagRewriteCell This can be used when a Cell has to change with addition/removal of one or more tags. - + ProcedureState POJO representing Procedure State - + RawCell An extended version of cell that gives more power to CPs - + RawCellBuilder Allows creating a cell with Tag An instance of this type can be acquired by using RegionCoprocessorEnvironment#getCellBuilder (for prod code) and RawCellBuilderFactory (for unit tests). - + RegionLoad Deprecated. As of release 2.0.0, this will be removed in HBase 3.0.0 @@ -768,32 +771,32 @@ service. - + RegionLocations Container for holding a list of HRegionLocation's that correspond to the same range. - + RegionMetrics Encapsulates per-region load metrics. - + RegionMetricsBuilder - + RetryImmediatelyException - + ScheduledChore ScheduledChore is a task performed on a period in hbase. - + ScheduledChore.ChoreServicer - + ServerLoad Deprecated. As of release 2.0.0, this will be removed in HBase 3.0.0 @@ -801,48 +804,48 @@ service. - + ServerMetrics This class is used for exporting current state of load on a RegionServer. - + ServerMetricsBuilder - + ServerName Name of a particular incarnation of an HBase Server. - + Size It is used to represent the size with different units. - + Size.Unit - + SizeCachedKeyValue This class is an extension to KeyValue where rowLen and keyLen are cached. - + SplitLogTask State of a
[31/51] [partial] hbase-site git commit: Published site at 1d25b60831b8cc8f7ad5fd366f1867de5c20d2f3.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/TableExistsException.html -- diff --git a/apidocs/org/apache/hadoop/hbase/TableExistsException.html b/apidocs/org/apache/hadoop/hbase/TableExistsException.html index c956448..c219c09 100644 --- a/apidocs/org/apache/hadoop/hbase/TableExistsException.html +++ b/apidocs/org/apache/hadoop/hbase/TableExistsException.html @@ -91,16 +91,16 @@ -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object -http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">java.lang.Throwable +https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">java.lang.Throwable -http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">java.lang.Exception +https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">java.lang.Exception -http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">java.io.IOException +https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">java.io.IOException org.apache.hadoop.hbase.HBaseIOException @@ -128,7 +128,7 @@ All Implemented Interfaces: -http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable +https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable @@ -163,7 +163,7 @@ extends -TableExistsException(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Strings) +TableExistsException(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Strings) Constructor @@ -183,15 +183,15 @@ extends -Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable -http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#addSuppressed-java.lang.Throwable-; title="class or interface in java.lang">addSuppressed, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#fillInStackTrace--; title="class or interface in java.lang">fillInStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getCause--; title="class or interface in java.lang">getCause, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getLocalizedMessage--; title="class or interface in java.lang">getLocalizedMessage, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getMessage--; title="class or interface in java.lang">getMessage, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getStackTrace--; title="class or inter face in java.lang">getStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getSuppressed--; title="class or interface in java.lang">getSuppressed, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#initCause-java.lang.Throwable-; title="class or interface in java.lang">initCause, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#printStackTrace--; title="class or interface in java.lang">printStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#printStackTrace-java.io.PrintStream-; title="class or interface in java.lang">printStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#printStackTrace-java.io.PrintWriter-; title="class or interface in java.lang">printStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lan g/Throwable.html?is-external=true#setStackTrace-java.lang.StackTraceElement:A-" title="class or interface in java.lang">setStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#toString--; title="class or interface in java.lang">toString +Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/org/apache/hadoop/hbase/client/MasterCallable.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/MasterCallable.html b/devapidocs/org/apache/hadoop/hbase/client/MasterCallable.html index bb9f461..3c505c9 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/MasterCallable.html +++ b/devapidocs/org/apache/hadoop/hbase/client/MasterCallable.html @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -abstract class MasterCallableV +abstract class MasterCallableV extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements RetryingCallableV, http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable A RetryingCallable for Master RPC operations. @@ -282,7 +282,7 @@ implements connection -protected finalClusterConnection connection +protected finalClusterConnection connection @@ -291,7 +291,7 @@ implements master -protectedMasterKeepAliveConnection master +protectedMasterKeepAliveConnection master @@ -300,7 +300,7 @@ implements rpcController -private finalHBaseRpcController rpcController +private finalHBaseRpcController rpcController @@ -317,7 +317,7 @@ implements MasterCallable -MasterCallable(Connectionconnection, +MasterCallable(Connectionconnection, RpcControllerFactoryrpcConnectionFactory) @@ -335,7 +335,7 @@ implements prepare -publicvoidprepare(booleanreload) +publicvoidprepare(booleanreload) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from interface:RetryingCallable Prepare by setting up any connections to servers, etc., ahead of call invocation. @@ -356,7 +356,7 @@ implements close -publicvoidclose() +publicvoidclose() throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Specified by: @@ -374,7 +374,7 @@ implements throwable -publicvoidthrowable(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablet, +publicvoidthrowable(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablet, booleanretrying) Description copied from interface:RetryingCallable Called when call throws an exception and we are going to retry; take action to @@ -396,7 +396,7 @@ implements getExceptionMessageAdditionalDetail -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetExceptionMessageAdditionalDetail() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetExceptionMessageAdditionalDetail() Specified by: getExceptionMessageAdditionalDetailin interfaceRetryingCallableV @@ -413,7 +413,7 @@ implements sleep -publiclongsleep(longpause, +publiclongsleep(longpause, inttries) Specified by: @@ -432,7 +432,7 @@ implements call -publicVcall(intcallTimeout) +publicVcall(intcallTimeout) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Override that changes the http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true#call--; title="class or interface in java.util.concurrent">Callable.call() Exception from http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception to http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException. It also does setup of an rpcController and calls through to the rpcCall() @@ -456,7 +456,7 @@ implements rpcCall -protected abstractVrpcCall() +protected abstractVrpcCall() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception Run the RPC call. Implement this method. To get at the rpcController that has been created and configured to make this rpc call, use getRpcController(). We are trying to contain @@ -475,7 +475,7 @@ implements getRpcController -HBaseRpcControllergetRpcController() +HBaseRpcControllergetRpcController() @@ -484,7 +484,7 @@ implements setPriority -voidsetPriority(intpriority) +voidsetPriority(intpriority) @@ -493,7 +493,7 @@ implements
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColByteBufferExtendedCell.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColByteBufferExtendedCell.html b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColByteBufferExtendedCell.html index d143ef8..4583895 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColByteBufferExtendedCell.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColByteBufferExtendedCell.html @@ -258,7 +258,7 @@ 250 251@Override 252public long heapSize() { -253 long sum = HEAP_SIZE_OVERHEAD + estimatedHeapSizeOf(cell); +253 long sum = HEAP_SIZE_OVERHEAD + estimatedSizeOfCell(cell); 254 if (this.tags != null) { 255sum += ClassSize.sizeOf(this.tags); 256 } @@ -454,7 +454,7 @@ 446 447@Override 448public long heapSize() { -449 long sum = HEAP_SIZE_OVERHEAD + estimatedHeapSizeOf(cell); +449 long sum = HEAP_SIZE_OVERHEAD + estimatedSizeOfCell(cell); 450 // this.tags is on heap byte[] 451 if (this.tags != null) { 452sum += ClassSize.sizeOf(this.tags); @@ -2791,192 +2791,193 @@ 2783 * {@link HeapSize} we call {@link HeapSize#heapSize()} so cell can give a correct value. In other 2784 * cases we just consider the bytes occupied by the cell components ie. row, CF, qualifier, 2785 * timestamp, type, value and tags. -2786 * @param cell -2787 * @return estimate of the heap space -2788 */ -2789 public static long estimatedHeapSizeOf(final Cell cell) { -2790if (cell instanceof HeapSize) { -2791 return ((HeapSize) cell).heapSize(); -2792} -2793// TODO: Add sizing of references that hold the row, family, etc., arrays. -2794return estimatedSerializedSizeOf(cell); -2795 } -2796 -2797 /** -2798 * This method exists just to encapsulate how we serialize keys. To be replaced by a factory that -2799 * we query to figure what the Cell implementation is and then, what serialization engine to use -2800 * and further, how to serialize the key for inclusion in hfile index. TODO. -2801 * @param cell -2802 * @return The key portion of the Cell serialized in the old-school KeyValue way or null if passed -2803 * a null codecell/code -2804 */ -2805 public static byte[] getCellKeySerializedAsKeyValueKey(final Cell cell) { -2806if (cell == null) return null; -2807byte[] b = new byte[KeyValueUtil.keyLength(cell)]; -2808KeyValueUtil.appendKeyTo(cell, b, 0); -2809return b; -2810 } -2811 -2812 /** -2813 * Create a Cell that is smaller than all other possible Cells for the given Cell's row. -2814 * @param cell -2815 * @return First possible Cell on passed Cell's row. -2816 */ -2817 public static Cell createFirstOnRow(final Cell cell) { -2818if (cell instanceof ByteBufferExtendedCell) { -2819 return new FirstOnRowByteBufferExtendedCell( -2820 ((ByteBufferExtendedCell) cell).getRowByteBuffer(), -2821 ((ByteBufferExtendedCell) cell).getRowPosition(), cell.getRowLength()); -2822} -2823return new FirstOnRowCell(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength()); -2824 } -2825 -2826 public static Cell createFirstOnRow(final byte[] row, int roffset, short rlength) { -2827return new FirstOnRowCell(row, roffset, rlength); -2828 } -2829 -2830 public static Cell createFirstOnRow(final byte[] row, final byte[] family, final byte[] col) { -2831return createFirstOnRow(row, 0, (short) row.length, family, 0, (byte) family.length, col, 0, -2832col.length); -2833 } -2834 -2835 public static Cell createFirstOnRow(final byte[] row, int roffset, short rlength, -2836 final byte[] family, int foffset, byte flength, final byte[] col, int coffset, int clength) { -2837return new FirstOnRowColCell(row, roffset, rlength, family, foffset, flength, col, coffset, -2838clength); -2839 } -2840 -2841 public static Cell createFirstOnRow(final byte[] row) { -2842return createFirstOnRow(row, 0, (short) row.length); -2843 } -2844 -2845 public static Cell createFirstOnRowFamily(Cell cell, byte[] fArray, int foff, int flen) { -2846if (cell instanceof ByteBufferExtendedCell) { -2847 return new FirstOnRowColByteBufferExtendedCell( -2848 ((ByteBufferExtendedCell) cell).getRowByteBuffer(), -2849 ((ByteBufferExtendedCell) cell).getRowPosition(), cell.getRowLength(), -2850 ByteBuffer.wrap(fArray), foff, (byte) flen, HConstants.EMPTY_BYTE_BUFFER, 0, 0); -2851} -2852return new FirstOnRowColCell(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength(), -2853fArray, foff, (byte) flen, HConstants.EMPTY_BYTE_ARRAY, 0, 0); -2854 }
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.SingleServerRequestRunnable.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.SingleServerRequestRunnable.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.SingleServerRequestRunnable.html index 00483be..c27b109 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.SingleServerRequestRunnable.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.SingleServerRequestRunnable.html @@ -396,888 +396,887 @@ 388new ConcurrentHashMapCancellableRegionServerCallable, Boolean()); 389this.asyncProcess = asyncProcess; 390this.errorsByServer = createServerErrorTracker(); -391this.errors = (asyncProcess.globalErrors != null) -392? asyncProcess.globalErrors : new BatchErrors(); -393this.operationTimeout = task.getOperationTimeout(); -394this.rpcTimeout = task.getRpcTimeout(); -395this.currentCallable = task.getCallable(); -396if (task.getCallable() == null) { -397 tracker = new RetryingTimeTracker().start(); -398} -399 } -400 -401 @VisibleForTesting -402 protected SetCancellableRegionServerCallable getCallsInProgress() { -403return callsInProgress; -404 } -405 -406 @VisibleForTesting -407 SingleServerRequestRunnable createSingleServerRequest(MultiAction multiAction, int numAttempt, ServerName server, -408 SetCancellableRegionServerCallable callsInProgress) { -409return new SingleServerRequestRunnable(multiAction, numAttempt, server, callsInProgress); -410 } -411 -412 /** -413 * Group a list of actions per region servers, and send them. -414 * -415 * @param currentActions - the list of row to submit -416 * @param numAttempt - the current numAttempt (first attempt is 1) -417 */ -418 void groupAndSendMultiAction(ListAction currentActions, int numAttempt) { -419MapServerName, MultiAction actionsByServer = new HashMap(); -420 -421boolean isReplica = false; -422ListAction unknownReplicaActions = null; -423for (Action action : currentActions) { -424 RegionLocations locs = findAllLocationsOrFail(action, true); -425 if (locs == null) continue; -426 boolean isReplicaAction = !RegionReplicaUtil.isDefaultReplica(action.getReplicaId()); -427 if (isReplica !isReplicaAction) { -428// This is the property of the current implementation, not a requirement. -429throw new AssertionError("Replica and non-replica actions in the same retry"); -430 } -431 isReplica = isReplicaAction; -432 HRegionLocation loc = locs.getRegionLocation(action.getReplicaId()); -433 if (loc == null || loc.getServerName() == null) { -434if (isReplica) { -435 if (unknownReplicaActions == null) { -436unknownReplicaActions = new ArrayList(1); -437 } -438 unknownReplicaActions.add(action); -439} else { -440 // TODO: relies on primary location always being fetched -441 manageLocationError(action, null); -442} -443 } else { -444byte[] regionName = loc.getRegionInfo().getRegionName(); -445 AsyncProcess.addAction(loc.getServerName(), regionName, action, actionsByServer, nonceGroup); -446 } -447} -448boolean doStartReplica = (numAttempt == 1 !isReplica hasAnyReplicaGets); -449boolean hasUnknown = unknownReplicaActions != null !unknownReplicaActions.isEmpty(); -450 -451if (!actionsByServer.isEmpty()) { -452 // If this is a first attempt to group and send, no replicas, we need replica thread. -453 sendMultiAction(actionsByServer, numAttempt, (doStartReplica !hasUnknown) -454 ? currentActions : null, numAttempt 1 !hasUnknown); -455} -456 -457if (hasUnknown) { -458 actionsByServer = new HashMap(); -459 for (Action action : unknownReplicaActions) { -460HRegionLocation loc = getReplicaLocationOrFail(action); -461if (loc == null) continue; -462byte[] regionName = loc.getRegionInfo().getRegionName(); -463 AsyncProcess.addAction(loc.getServerName(), regionName, action, actionsByServer, nonceGroup); -464 } -465 if (!actionsByServer.isEmpty()) { -466sendMultiAction( -467actionsByServer, numAttempt, doStartReplica ? currentActions : null, true); -468 } -469} -470 } -471 -472 private HRegionLocation getReplicaLocationOrFail(Action action) { -473// We are going to try get location once again. For each action, we'll do it once -474// from cache, because the previous calls in the loop might populate it. -475int replicaId = action.getReplicaId(); -476RegionLocations locs =
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html index c7fba62..eca6413 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html @@ -127,30 +127,28 @@ the order they are declared. -private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -RawAsyncHBaseAdmin.compact(TableNametableName, +private void +HBaseAdmin.compact(TableNametableName, byte[]columnFamily, booleanmajor, CompactTypecompactType) -Compact column family of a table, Asynchronous operation even if CompletableFuture.get() +Compact a table. -private void -HBaseAdmin.compact(TableNametableName, +private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +RawAsyncHBaseAdmin.compact(TableNametableName, byte[]columnFamily, booleanmajor, CompactTypecompactType) -Compact a table. +Compact column family of a table, Asynchronous operation even if CompletableFuture.get() http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncAdmin.compact(TableNametableName, +AsyncHBaseAdmin.compact(TableNametableName, byte[]columnFamily, - CompactTypecompactType) -Compact a column family within a table. - + CompactTypecompactType) void @@ -161,14 +159,16 @@ the order they are declared. -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -RawAsyncHBaseAdmin.compact(TableNametableName, +void +HBaseAdmin.compact(TableNametableName, byte[]columnFamily, - CompactTypecompactType) + CompactTypecompactType) +Compact a column family within a table. + -void -HBaseAdmin.compact(TableNametableName, +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +AsyncAdmin.compact(TableNametableName, byte[]columnFamily, CompactTypecompactType) Compact a column family within a table. @@ -176,16 +176,14 @@ the order they are declared. http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncHBaseAdmin.compact(TableNametableName, +RawAsyncHBaseAdmin.compact(TableNametableName, byte[]columnFamily, CompactTypecompactType) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncAdmin.compact(TableNametableName, - CompactTypecompactType) -Compact a table. - +AsyncHBaseAdmin.compact(TableNametableName, + CompactTypecompactType) void @@ -195,28 +193,28 @@ the order they are declared. -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -RawAsyncHBaseAdmin.compact(TableNametableName, - CompactTypecompactType) - - void HBaseAdmin.compact(TableNametableName, CompactTypecompactType) Compact a table. +
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/class-use/Size.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Size.html b/devapidocs/org/apache/hadoop/hbase/class-use/Size.html index 6c0e3b8..20b7674 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/Size.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Size.html @@ -189,130 +189,130 @@ Size -RegionLoad.getBloomFilterSize() -Deprecated. - - - -Size RegionMetrics.getBloomFilterSize() - -Size -RegionMetricsBuilder.RegionMetricsImpl.getBloomFilterSize() - Size -ServerLoad.getMaxHeapSize() +RegionLoad.getBloomFilterSize() Deprecated. Size -ServerMetrics.getMaxHeapSize() +RegionMetricsBuilder.RegionMetricsImpl.getBloomFilterSize() Size -ServerMetricsBuilder.ServerMetricsImpl.getMaxHeapSize() +ServerMetrics.getMaxHeapSize() Size -RegionLoad.getMemStoreSize() +ServerLoad.getMaxHeapSize() Deprecated. Size -RegionMetrics.getMemStoreSize() +ServerMetricsBuilder.ServerMetricsImpl.getMaxHeapSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getMemStoreSize() +RegionMetrics.getMemStoreSize() Size -RegionLoad.getStoreFileIndexSize() +RegionLoad.getMemStoreSize() Deprecated. Size +RegionMetricsBuilder.RegionMetricsImpl.getMemStoreSize() + + +Size RegionMetrics.getStoreFileIndexSize() TODO: why we pass the same value to different counters? Currently, the value from getStoreFileIndexSize() is same with getStoreFileRootLevelIndexSize() see HRegionServer#createRegionLoad. - -Size -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileIndexSize() - Size -RegionLoad.getStoreFileRootLevelIndexSize() +RegionLoad.getStoreFileIndexSize() Deprecated. Size -RegionMetrics.getStoreFileRootLevelIndexSize() +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileIndexSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileRootLevelIndexSize() +RegionMetrics.getStoreFileRootLevelIndexSize() Size -RegionLoad.getStoreFileSize() +RegionLoad.getStoreFileRootLevelIndexSize() Deprecated. Size -RegionMetrics.getStoreFileSize() +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileRootLevelIndexSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileSize() +RegionMetrics.getStoreFileSize() Size -RegionLoad.getStoreFileUncompressedDataIndexSize() +RegionLoad.getStoreFileSize() Deprecated. Size -RegionMetrics.getStoreFileUncompressedDataIndexSize() +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileUncompressedDataIndexSize() +RegionMetrics.getStoreFileUncompressedDataIndexSize() Size -RegionLoad.getUncompressedStoreFileSize() +RegionLoad.getStoreFileUncompressedDataIndexSize() Deprecated. Size -RegionMetrics.getUncompressedStoreFileSize() +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileUncompressedDataIndexSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getUncompressedStoreFileSize() +RegionMetrics.getUncompressedStoreFileSize() Size -ServerLoad.getUsedHeapSize() +RegionLoad.getUncompressedStoreFileSize() Deprecated. Size +RegionMetricsBuilder.RegionMetricsImpl.getUncompressedStoreFileSize() + + +Size ServerMetrics.getUsedHeapSize() + +Size +ServerLoad.getUsedHeapSize() +Deprecated. + + Size ServerMetricsBuilder.ServerMetricsImpl.getUsedHeapSize() http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html index 320e7bc..7f3e934 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html @@ -122,11 +122,11 @@ TableDescriptors -HMaster.getTableDescriptors() +MasterServices.getTableDescriptors() TableDescriptors -MasterServices.getTableDescriptors() +HMaster.getTableDescriptors()
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html index 7eb7661..1d50582 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html @@ -104,13 +104,13 @@ void -MasterServices.checkTableModifiable(TableNametableName) -Check table is modifiable; i.e. - +HMaster.checkTableModifiable(TableNametableName) void -HMaster.checkTableModifiable(TableNametableName) +MasterServices.checkTableModifiable(TableNametableName) +Check table is modifiable; i.e. + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html index b1a475b..a32bb19 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html @@ -170,13 +170,13 @@ void -MasterServices.checkTableModifiable(TableNametableName) -Check table is modifiable; i.e. - +HMaster.checkTableModifiable(TableNametableName) void -HMaster.checkTableModifiable(TableNametableName) +MasterServices.checkTableModifiable(TableNametableName) +Check table is modifiable; i.e. + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html index 90f6967..594e74b 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html @@ -243,10 +243,6 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTag -PrivateCellUtil.getTags(Cellcell) - - -static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTag CellUtil.getTags(Cellcell) Deprecated. As of release 2.0.0, this will be removed in HBase 3.0.0. @@ -254,6 +250,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. + +static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTag +PrivateCellUtil.getTags(Cellcell) + static http://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true; title="class or interface in java.util">IteratorTag CellUtil.tagsIterator(byte[]tags, @@ -395,11 +395,6 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. static Cell -PrivateCellUtil.createCell(Cellcell, - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) - - -static Cell CellUtil.createCell(Cellcell, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) Deprecated. @@ -407,6 +402,11 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. + +static Cell +PrivateCellUtil.createCell(Cellcell, + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) + static byte[] TagUtil.fromList(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) @@ -415,17 +415,17 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -ExtendedCellBuilder -ExtendedCellBuilderImpl.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) - - RawCellBuilder RawCellBuilder.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) - + ExtendedCellBuilder ExtendedCellBuilder.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) + +ExtendedCellBuilder +ExtendedCellBuilderImpl.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/NamespaceDescriptor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/NamespaceDescriptor.html b/devapidocs/org/apache/hadoop/hbase/class-use/NamespaceDescriptor.html index b05384b..f9cacbc 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/NamespaceDescriptor.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/NamespaceDescriptor.html @@ -270,32 +270,32 @@ http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureNamespaceDescriptor -AsyncHBaseAdmin.getNamespaceDescriptor(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) - - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureNamespaceDescriptor AsyncAdmin.getNamespaceDescriptor(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Get a namespace descriptor by name - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureNamespaceDescriptor RawAsyncHBaseAdmin.getNamespaceDescriptor(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListNamespaceDescriptor -AsyncHBaseAdmin.listNamespaceDescriptors() - +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureNamespaceDescriptor +AsyncHBaseAdmin.getNamespaceDescriptor(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) + + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListNamespaceDescriptor AsyncAdmin.listNamespaceDescriptors() List available namespace descriptors - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListNamespaceDescriptor RawAsyncHBaseAdmin.listNamespaceDescriptors() + +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListNamespaceDescriptor +AsyncHBaseAdmin.listNamespaceDescriptors() + @@ -307,7 +307,9 @@ http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncHBaseAdmin.createNamespace(NamespaceDescriptordescriptor) +AsyncAdmin.createNamespace(NamespaceDescriptordescriptor) +Create a new namespace. + void @@ -316,18 +318,16 @@ -void -HBaseAdmin.createNamespace(NamespaceDescriptordescriptor) +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +RawAsyncHBaseAdmin.createNamespace(NamespaceDescriptordescriptor) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncAdmin.createNamespace(NamespaceDescriptordescriptor) -Create a new namespace. - +void +HBaseAdmin.createNamespace(NamespaceDescriptordescriptor)
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html index 7eb7661..1d50582 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html @@ -104,13 +104,13 @@ void -MasterServices.checkTableModifiable(TableNametableName) -Check table is modifiable; i.e. - +HMaster.checkTableModifiable(TableNametableName) void -HMaster.checkTableModifiable(TableNametableName) +MasterServices.checkTableModifiable(TableNametableName) +Check table is modifiable; i.e. + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html index 818c4ca..5eebc7c 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html @@ -157,13 +157,13 @@ void -MasterServices.checkTableModifiable(TableNametableName) -Check table is modifiable; i.e. - +HMaster.checkTableModifiable(TableNametableName) void -HMaster.checkTableModifiable(TableNametableName) +MasterServices.checkTableModifiable(TableNametableName) +Check table is modifiable; i.e. + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html index 90f6967..594e74b 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html @@ -243,10 +243,6 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTag -PrivateCellUtil.getTags(Cellcell) - - -static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTag CellUtil.getTags(Cellcell) Deprecated. As of release 2.0.0, this will be removed in HBase 3.0.0. @@ -254,6 +250,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. + +static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTag +PrivateCellUtil.getTags(Cellcell) + static http://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true; title="class or interface in java.util">IteratorTag CellUtil.tagsIterator(byte[]tags, @@ -395,11 +395,6 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. static Cell -PrivateCellUtil.createCell(Cellcell, - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) - - -static Cell CellUtil.createCell(Cellcell, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) Deprecated. @@ -407,6 +402,11 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. + +static Cell +PrivateCellUtil.createCell(Cellcell, + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) + static byte[] TagUtil.fromList(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) @@ -415,17 +415,17 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -ExtendedCellBuilder -ExtendedCellBuilderImpl.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) - - RawCellBuilder RawCellBuilder.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) - + ExtendedCellBuilder ExtendedCellBuilder.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) + +ExtendedCellBuilder +ExtendedCellBuilderImpl.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html index ded834e..9c4acf5 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html @@ -330,8 +330,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncHBaseAdmin.addColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) +AsyncAdmin.addColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) +Add a column family to an existing table. + void @@ -341,20 +343,18 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -void -HBaseAdmin.addColumnFamily(TableNametableName, +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +RawAsyncHBaseAdmin.addColumnFamily(TableNametableName, ColumnFamilyDescriptorcolumnFamily) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncAdmin.addColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) -Add a column family to an existing table. - +void +HBaseAdmin.addColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -RawAsyncHBaseAdmin.addColumnFamily(TableNametableName, +AsyncHBaseAdmin.addColumnFamily(TableNametableName, ColumnFamilyDescriptorcolumnFamily) @@ -396,8 +396,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncHBaseAdmin.modifyColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) +AsyncAdmin.modifyColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) +Modify an existing column family on a table. + void @@ -407,20 +409,18 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -void -HBaseAdmin.modifyColumnFamily(TableNametableName, +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +RawAsyncHBaseAdmin.modifyColumnFamily(TableNametableName, ColumnFamilyDescriptorcolumnFamily) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncAdmin.modifyColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) -Modify an existing column family on a table. - +void +HBaseAdmin.modifyColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALActionListener.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALActionListener.html b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALActionListener.html new file mode 100644 index 000..c32f3bb --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALActionListener.html @@ -0,0 +1,439 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +ReplicationSourceWALActionListener (Apache HBase 3.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":9,"i3":10}; +var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.replication.regionserver +Class ReplicationSourceWALActionListener + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceWALActionListener + + + + + + + +All Implemented Interfaces: +WALActionsListener + + + +@InterfaceAudience.Private +class ReplicationSourceWALActionListener +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +implements WALActionsListener +Used to receive new wals. + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +private org.apache.hadoop.conf.Configuration +conf + + +private ReplicationSourceManager +manager + + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +ReplicationSourceWALActionListener(org.apache.hadoop.conf.Configurationconf, + ReplicationSourceManagermanager) + + + + + + + + + +Method Summary + +All MethodsStatic MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +void +postLogRoll(org.apache.hadoop.fs.PatholdPath, + org.apache.hadoop.fs.PathnewPath) +The WAL has been rolled. + + + +void +preLogRoll(org.apache.hadoop.fs.PatholdPath, + org.apache.hadoop.fs.PathnewPath) +The WAL is going to be rolled. + + + +(package private) static void +scopeWALEdits(WALKeylogKey, + WALEditlogEdit, + org.apache.hadoop.conf.Configurationconf) +Utility method used to set the correct scopes on each log key. + + + +void +visitLogEntryBeforeWrite(WALKeylogKey, +WALEditlogEdit) + + + + + + +Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify, http://docs.oracle.com/javase/8/docs/api/java/lang /Object.html?is-external=true#notifyAll--" title="class or interface in java.lang">notifyAll, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--; title="class or interface in java.lang">toString,
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/regionserver/wal/class-use/WALActionsListener.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/class-use/WALActionsListener.html b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/class-use/WALActionsListener.html index 8a18a7d..df45e2f 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/class-use/WALActionsListener.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/class-use/WALActionsListener.html @@ -83,17 +83,9 @@ -org.apache.hadoop.hbase.regionserver - - - org.apache.hadoop.hbase.regionserver.wal - -org.apache.hadoop.hbase.replication.regionserver - - org.apache.hadoop.hbase.wal @@ -103,26 +95,6 @@ - - - -Uses of WALActionsListener in org.apache.hadoop.hbase.regionserver - -Methods in org.apache.hadoop.hbase.regionserver that return WALActionsListener - -Modifier and Type -Method and Description - - - -WALActionsListener -ReplicationSourceService.getWALActionsListener() -Returns a WALObserver for the service. - - - - - @@ -220,39 +192,6 @@ - - - -Uses of WALActionsListener in org.apache.hadoop.hbase.replication.regionserver - -Classes in org.apache.hadoop.hbase.replication.regionserver that implement WALActionsListener - -Modifier and Type -Class and Description - - - -class -Replication -Gateway to Replication. - - - - - -Methods in org.apache.hadoop.hbase.replication.regionserver that return WALActionsListener - -Modifier and Type -Method and Description - - - -WALActionsListener -Replication.getWALActionsListener() - - - - @@ -287,77 +226,41 @@ void -DisabledWALProvider.DisabledWAL.registerWALActionsListener(WALActionsListenerlistener) +AbstractFSWALProvider.addWALActionsListener(WALActionsListenerlistener) void -WAL.registerWALActionsListener(WALActionsListenerlistener) -Registers WALActionsListener - +RegionGroupingProvider.addWALActionsListener(WALActionsListenerlistener) -boolean -DisabledWALProvider.DisabledWAL.unregisterWALActionsListener(WALActionsListenerlistener) - - -boolean -WAL.unregisterWALActionsListener(WALActionsListenerlistener) -Unregisters WALActionsListener +void +WALProvider.addWALActionsListener(WALActionsListenerlistener) +Add a WALActionsListener. - - - -Method parameters in org.apache.hadoop.hbase.wal with type arguments of type WALActionsListener - -Modifier and Type -Method and Description - - - -(package private) WALProvider -WALFactory.createProvider(http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends WALProviderclazz, - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListWALActionsListenerlisteners, - http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringproviderId) - -(package private) WALProvider -WALFactory.getProvider(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringkey, - http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringdefaultValue, - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListWALActionsListenerlisteners, - http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringproviderId) -instantiate a provider from a config property. - +void +DisabledWALProvider.addWALActionsListener(WALActionsListenerlistener) void -AbstractFSWALProvider.init(WALFactoryfactory, -org.apache.hadoop.conf.Configurationconf, -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListWALActionsListenerlisteners, -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringproviderId) +DisabledWALProvider.DisabledWAL.registerWALActionsListener(WALActionsListenerlistener) void -RegionGroupingProvider.init(WALFactoryfactory, -org.apache.hadoop.conf.Configurationconf, -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListWALActionsListenerlisteners, -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringproviderId) +WAL.registerWALActionsListener(WALActionsListenerlistener) +Registers WALActionsListener + -void -WALProvider.init(WALFactoryfactory, -org.apache.hadoop.conf.Configurationconf, -
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/TestJMXConnectorServer.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/TestJMXConnectorServer.html b/testdevapidocs/org/apache/hadoop/hbase/TestJMXConnectorServer.html index 3148d86..7f3ff92 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/TestJMXConnectorServer.html +++ b/testdevapidocs/org/apache/hadoop/hbase/TestJMXConnectorServer.html @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab"; -public class TestJMXConnectorServer +public class TestJMXConnectorServer extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object Test case for JMX Connector Server. @@ -154,22 +154,26 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? admin +static HBaseClassTestRule +CLASS_RULE + + private static org.apache.hadoop.conf.Configuration conf - + (package private) static boolean hasAccess - + private static org.slf4j.Logger LOG - + private static int rmiRegistryPort - + private static HBaseTestingUtility UTIL @@ -254,13 +258,22 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? Field Detail + + + + + +CLASS_RULE +public static finalHBaseClassTestRule CLASS_RULE + + LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -269,7 +282,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? UTIL -private staticHBaseTestingUtility UTIL +private staticHBaseTestingUtility UTIL @@ -278,7 +291,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? conf -private staticorg.apache.hadoop.conf.Configuration conf +private staticorg.apache.hadoop.conf.Configuration conf @@ -287,7 +300,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? admin -private staticorg.apache.hadoop.hbase.client.Admin admin +private staticorg.apache.hadoop.hbase.client.Admin admin @@ -296,7 +309,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? rmiRegistryPort -private staticint rmiRegistryPort +private staticint rmiRegistryPort @@ -305,7 +318,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? hasAccess -staticboolean hasAccess +staticboolean hasAccess @@ -322,7 +335,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? TestJMXConnectorServer -publicTestJMXConnectorServer() +publicTestJMXConnectorServer() @@ -339,7 +352,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? setUp -publicvoidsetUp() +publicvoidsetUp() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception Throws: @@ -353,7 +366,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? tearDown -publicvoidtearDown() +publicvoidtearDown() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception Throws: @@ -367,7 +380,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? testHMConnectorServerWhenStopMaster -publicvoidtestHMConnectorServerWhenStopMaster() +publicvoidtestHMConnectorServerWhenStopMaster() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception This tests to validate the HMaster's ConnectorServer after unauthorised stopMaster call. @@ -382,7 +395,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? testRSConnectorServerWhenStopRegionServer -publicvoidtestRSConnectorServerWhenStopRegionServer() +publicvoidtestRSConnectorServerWhenStopRegionServer() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception This tests to validate the RegionServer's ConnectorServer after unauthorised stopRegionServer call. @@ -398,7 +411,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? testHMConnectorServerWhenShutdownCluster -publicvoidtestHMConnectorServerWhenShutdownCluster() +publicvoidtestHMConnectorServerWhenShutdownCluster() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception This tests to validate the HMaster's ConnectorServer after unauthorised shutdown call.
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupAdminImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupAdminImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupAdminImpl.html index 62efc6b..12f4f50 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupAdminImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupAdminImpl.html @@ -52,61 +52,61 @@ 044import org.apache.hadoop.hbase.backup.RestoreRequest; 045import org.apache.hadoop.hbase.backup.util.BackupSet; 046import org.apache.hadoop.hbase.backup.util.BackupUtils; -047import org.apache.yetus.audience.InterfaceAudience; -048import org.slf4j.Logger; -049import org.slf4j.LoggerFactory; -050import org.apache.hadoop.hbase.client.Admin; -051import org.apache.hadoop.hbase.client.Connection; -052import org.apache.hbase.thirdparty.com.google.common.collect.Lists; -053import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -054 -055@InterfaceAudience.Private -056public class BackupAdminImpl implements BackupAdmin { -057 public final static String CHECK_OK = "Checking backup images: OK"; -058 public final static String CHECK_FAILED = -059 "Checking backup images: Failed. Some dependencies are missing for restore"; -060 private static final Logger LOG = LoggerFactory.getLogger(BackupAdminImpl.class); -061 -062 private final Connection conn; -063 -064 public BackupAdminImpl(Connection conn) { -065this.conn = conn; -066 } -067 -068 @Override -069 public void close() throws IOException { -070 } -071 -072 @Override -073 public BackupInfo getBackupInfo(String backupId) throws IOException { -074BackupInfo backupInfo = null; -075try (final BackupSystemTable table = new BackupSystemTable(conn)) { -076 if (backupId == null) { -077ArrayListBackupInfo recentSessions = table.getBackupInfos(BackupState.RUNNING); -078if (recentSessions.isEmpty()) { -079 LOG.warn("No ongoing sessions found."); -080 return null; -081} -082// else show status for ongoing session -083// must be one maximum -084return recentSessions.get(0); -085 } else { -086backupInfo = table.readBackupInfo(backupId); -087return backupInfo; -088 } -089} -090 } -091 -092 @Override -093 public int deleteBackups(String[] backupIds) throws IOException { -094 -095int totalDeleted = 0; -096MapString, HashSetTableName allTablesMap = new HashMapString, HashSetTableName(); -097 -098boolean deleteSessionStarted = false; -099boolean snapshotDone = false; -100try (final BackupSystemTable sysTable = new BackupSystemTable(conn)) { -101 +047import org.apache.hadoop.hbase.client.Admin; +048import org.apache.hadoop.hbase.client.Connection; +049import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +050import org.apache.yetus.audience.InterfaceAudience; +051import org.slf4j.Logger; +052import org.slf4j.LoggerFactory; +053 +054import org.apache.hbase.thirdparty.com.google.common.collect.Lists; +055 +056@InterfaceAudience.Private +057public class BackupAdminImpl implements BackupAdmin { +058 public final static String CHECK_OK = "Checking backup images: OK"; +059 public final static String CHECK_FAILED = +060 "Checking backup images: Failed. Some dependencies are missing for restore"; +061 private static final Logger LOG = LoggerFactory.getLogger(BackupAdminImpl.class); +062 +063 private final Connection conn; +064 +065 public BackupAdminImpl(Connection conn) { +066this.conn = conn; +067 } +068 +069 @Override +070 public void close() { +071 } +072 +073 @Override +074 public BackupInfo getBackupInfo(String backupId) throws IOException { +075BackupInfo backupInfo; +076try (final BackupSystemTable table = new BackupSystemTable(conn)) { +077 if (backupId == null) { +078ArrayListBackupInfo recentSessions = table.getBackupInfos(BackupState.RUNNING); +079if (recentSessions.isEmpty()) { +080 LOG.warn("No ongoing sessions found."); +081 return null; +082} +083// else show status for ongoing session +084// must be one maximum +085return recentSessions.get(0); +086 } else { +087backupInfo = table.readBackupInfo(backupId); +088return backupInfo; +089 } +090} +091 } +092 +093 @Override +094 public int deleteBackups(String[] backupIds) throws IOException { +095 +096int totalDeleted = 0; +097MapString, HashSetTableName allTablesMap = new HashMap(); +098 +099boolean deleteSessionStarted; +100boolean snapshotDone; +101try (final BackupSystemTable sysTable = new BackupSystemTable(conn)) { 102 // Step 1: Make sure there is no active session 103
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html index ca8be5e..b8e6dfa 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html @@ -6398,514 +6398,514 @@ 6390 int initialBatchProgress = scannerContext.getBatchProgress(); 6391 long initialSizeProgress = scannerContext.getDataSizeProgress(); 6392 long initialHeapSizeProgress = scannerContext.getHeapSizeProgress(); -6393 long initialTimeProgress = scannerContext.getTimeProgress(); -6394 -6395 // The loop here is used only when at some point during the next we determine -6396 // that due to effects of filters or otherwise, we have an empty row in the result. -6397 // Then we loop and try again. Otherwise, we must get out on the first iteration via return, -6398 // "true" if there's more data to read, "false" if there isn't (storeHeap is at a stop row, -6399 // and joinedHeap has no more data to read for the last row (if set, joinedContinuationRow). -6400 while (true) { -6401// Starting to scan a new row. Reset the scanner progress according to whether or not -6402// progress should be kept. -6403if (scannerContext.getKeepProgress()) { -6404 // Progress should be kept. Reset to initial values seen at start of method invocation. -6405 scannerContext.setProgress(initialBatchProgress, initialSizeProgress, -6406 initialHeapSizeProgress, initialTimeProgress); -6407} else { -6408 scannerContext.clearProgress(); -6409} -6410if (rpcCall.isPresent()) { -6411 // If a user specifies a too-restrictive or too-slow scanner, the -6412 // client might time out and disconnect while the server side -6413 // is still processing the request. We should abort aggressively -6414 // in that case. -6415 long afterTime = rpcCall.get().disconnectSince(); -6416 if (afterTime = 0) { -6417throw new CallerDisconnectedException( -6418"Aborting on region " + getRegionInfo().getRegionNameAsString() + ", call " + -6419this + " after " + afterTime + " ms, since " + -6420"caller disconnected"); -6421 } -6422} -6423 -6424// Let's see what we have in the storeHeap. -6425Cell current = this.storeHeap.peek(); -6426 -6427boolean shouldStop = shouldStop(current); -6428// When has filter row is true it means that the all the cells for a particular row must be -6429// read before a filtering decision can be made. This means that filters where hasFilterRow -6430// run the risk of enLongAddering out of memory errors in the case that they are applied to a -6431// table that has very large rows. -6432boolean hasFilterRow = this.filter != null this.filter.hasFilterRow(); -6433 -6434// If filter#hasFilterRow is true, partial results are not allowed since allowing them -6435// would prevent the filters from being evaluated. Thus, if it is true, change the -6436// scope of any limits that could potentially create partial results to -6437// LimitScope.BETWEEN_ROWS so that those limits are not reached mid-row -6438if (hasFilterRow) { -6439 if (LOG.isTraceEnabled()) { -6440 LOG.trace("filter#hasFilterRow is true which prevents partial results from being " -6441+ " formed. Changing scope of limits that may create partials"); -6442 } -6443 scannerContext.setSizeLimitScope(LimitScope.BETWEEN_ROWS); -6444 scannerContext.setTimeLimitScope(LimitScope.BETWEEN_ROWS); -6445} -6446 -6447// Check if we were getting data from the joinedHeap and hit the limit. -6448// If not, then it's main path - getting results from storeHeap. -6449if (joinedContinuationRow == null) { -6450 // First, check if we are at a stop row. If so, there are no more results. -6451 if (shouldStop) { -6452if (hasFilterRow) { -6453 filter.filterRowCells(results); -6454} -6455return scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues(); -6456 } -6457 -6458 // Check if rowkey filter wants to exclude this row. If so, loop to next. -6459 // Technically, if we hit limits before on this row, we don't need this call. -6460 if
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/coprocessor/ReadOnlyConfiguration.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/ReadOnlyConfiguration.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/ReadOnlyConfiguration.html index 815025b..4c05ae9 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/ReadOnlyConfiguration.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/ReadOnlyConfiguration.html @@ -255,8 +255,8 @@ extends org.apache.hadoop.conf.Configuration http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class?[] -getClasses(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname, - http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class?[]defaultValue) +getClasses(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname, + http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class?...defaultValue) http://docs.oracle.com/javase/8/docs/api/java/lang/ClassLoader.html?is-external=true; title="class or interface in java.lang">ClassLoader @@ -743,7 +743,7 @@ extends org.apache.hadoop.conf.Configuration setAllowNullValueProperties -publicvoidsetAllowNullValueProperties(booleanval) +publicvoidsetAllowNullValueProperties(booleanval) Overrides: setAllowNullValuePropertiesin classorg.apache.hadoop.conf.Configuration @@ -756,7 +756,7 @@ extends org.apache.hadoop.conf.Configuration getTrimmed -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetTrimmed(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetTrimmed(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Overrides: getTrimmedin classorg.apache.hadoop.conf.Configuration @@ -769,7 +769,7 @@ extends org.apache.hadoop.conf.Configuration getTrimmed -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetTrimmed(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname, +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetTrimmed(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringdefaultValue) Overrides: @@ -783,7 +783,7 @@ extends org.apache.hadoop.conf.Configuration getRaw -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetRaw(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetRaw(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Overrides: getRawin classorg.apache.hadoop.conf.Configuration @@ -796,7 +796,7 @@ extends org.apache.hadoop.conf.Configuration set -publicvoidset(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname, +publicvoidset(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringvalue) Overrides: @@ -810,7 +810,7 @@ extends org.apache.hadoop.conf.Configuration set -publicvoidset(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname, +publicvoidset(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname,
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html index b50a65f..7271567 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html @@ -1718,312 +1718,314 @@ 1710 1711public WorkerThread(final ThreadGroup group) { 1712 super(group, "ProcExecWrkr-" + workerId.incrementAndGet()); -1713} -1714 -1715@Override -1716public void sendStopSignal() { -1717 scheduler.signalAll(); -1718} -1719 -1720@Override -1721public void run() { -1722 long lastUpdate = EnvironmentEdgeManager.currentTime(); -1723 try { -1724while (isRunning() keepAlive(lastUpdate)) { -1725 this.activeProcedure = scheduler.poll(keepAliveTime, TimeUnit.MILLISECONDS); -1726 if (this.activeProcedure == null) continue; -1727 int activeCount = activeExecutorCount.incrementAndGet(); -1728 int runningCount = store.setRunningProcedureCount(activeCount); -1729 if (LOG.isTraceEnabled()) { -1730LOG.trace("Execute pid=" + this.activeProcedure.getProcId() + -1731" runningCount=" + runningCount + ", activeCount=" + activeCount); -1732 } -1733 executionStartTime.set(EnvironmentEdgeManager.currentTime()); -1734 try { -1735 executeProcedure(this.activeProcedure); -1736 } catch (AssertionError e) { -1737LOG.info("ASSERT pid=" + this.activeProcedure.getProcId(), e); -1738throw e; -1739 } finally { -1740activeCount = activeExecutorCount.decrementAndGet(); -1741runningCount = store.setRunningProcedureCount(activeCount); -1742if (LOG.isTraceEnabled()) { -1743 LOG.trace("Halt pid=" + this.activeProcedure.getProcId() + -1744 " runningCount=" + runningCount + ", activeCount=" + activeCount); -1745} -1746this.activeProcedure = null; -1747lastUpdate = EnvironmentEdgeManager.currentTime(); -1748 executionStartTime.set(Long.MAX_VALUE); -1749 } -1750} -1751 } catch (Throwable t) { -1752LOG.warn("Worker terminating UNNATURALLY " + this.activeProcedure, t); -1753 } finally { -1754LOG.debug("Worker terminated."); -1755 } -1756 workerThreads.remove(this); -1757} -1758 -1759@Override -1760public String toString() { -1761 Procedure? p = this.activeProcedure; -1762 return getName() + "(pid=" + (p == null? Procedure.NO_PROC_ID: p.getProcId() + ")"); -1763} -1764 -1765/** -1766 * @return the time since the current procedure is running -1767 */ -1768public long getCurrentRunTime() { -1769 return EnvironmentEdgeManager.currentTime() - executionStartTime.get(); -1770} -1771 -1772private boolean keepAlive(final long lastUpdate) { -1773 if (workerThreads.size() = corePoolSize) return true; -1774 return (EnvironmentEdgeManager.currentTime() - lastUpdate) keepAliveTime; -1775} -1776 } -1777 -1778 /** -1779 * Runs task on a period such as check for stuck workers. -1780 * @see InlineChore -1781 */ -1782 private final class TimeoutExecutorThread extends StoppableThread { -1783private final DelayQueueDelayedWithTimeout queue = new DelayQueue(); -1784 -1785public TimeoutExecutorThread(final ThreadGroup group) { -1786 super(group, "ProcExecTimeout"); -1787} -1788 -1789@Override -1790public void sendStopSignal() { -1791 queue.add(DelayedUtil.DELAYED_POISON); -1792} -1793 -1794@Override -1795public void run() { -1796 final boolean traceEnabled = LOG.isTraceEnabled(); -1797 while (isRunning()) { -1798final DelayedWithTimeout task = DelayedUtil.takeWithoutInterrupt(queue); -1799if (task == null || task == DelayedUtil.DELAYED_POISON) { -1800 // the executor may be shutting down, -1801 // and the task is just the shutdown request -1802 continue; -1803} -1804 -1805if (traceEnabled) { -1806 LOG.trace("Executing " + task); -1807} -1808 -1809// execute the task -1810if (task instanceof InlineChore) { -1811 execInlineChore((InlineChore)task); -1812} else if (task instanceof DelayedProcedure) { -1813 execDelayedProcedure((DelayedProcedure)task); -1814} else { -1815 LOG.error("CODE-BUG unknown timeout task type " + task);
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.Task.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.Task.html b/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.Task.html index 5b66298..ea864e9 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.Task.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.Task.html @@ -39,319 +39,329 @@ 031import java.util.concurrent.atomic.AtomicBoolean; 032 033import org.apache.hadoop.conf.Configuration; -034import org.apache.yetus.audience.InterfaceAudience; -035import org.apache.zookeeper.KeeperException; -036import org.apache.zookeeper.KeeperException.Code; -037import org.apache.zookeeper.ZooKeeper; -038import org.apache.zookeeper.data.Stat; -039import org.slf4j.Logger; -040import org.slf4j.LoggerFactory; -041import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -042 -043/** -044 * A very simple read only zookeeper implementation without watcher support. -045 */ -046@InterfaceAudience.Private -047public final class ReadOnlyZKClient implements Closeable { -048 -049 private static final Logger LOG = LoggerFactory.getLogger(ReadOnlyZKClient.class); -050 -051 public static final String RECOVERY_RETRY = "zookeeper.recovery.retry"; -052 -053 private static final int DEFAULT_RECOVERY_RETRY = 30; -054 -055 public static final String RECOVERY_RETRY_INTERVAL_MILLIS = -056 "zookeeper.recovery.retry.intervalmill"; -057 -058 private static final int DEFAULT_RECOVERY_RETRY_INTERVAL_MILLIS = 1000; -059 -060 public static final String KEEPALIVE_MILLIS = "zookeeper.keep-alive.time"; -061 -062 private static final int DEFAULT_KEEPALIVE_MILLIS = 6; -063 -064 private static final EnumSetCode FAIL_FAST_CODES = EnumSet.of(Code.NOAUTH, Code.AUTHFAILED); -065 -066 private final String connectString; -067 -068 private final int sessionTimeoutMs; -069 -070 private final int maxRetries; -071 -072 private final int retryIntervalMs; -073 -074 private final int keepAliveTimeMs; -075 -076 private static abstract class Task implements Delayed { -077 -078protected long time = System.nanoTime(); -079 -080public boolean needZk() { -081 return false; -082} -083 -084public void exec(ZooKeeper zk) { -085} -086 -087public void connectFailed(IOException e) { -088} -089 -090public void closed(IOException e) { -091} -092 -093@Override -094public int compareTo(Delayed o) { -095 Task that = (Task) o; -096 int c = Long.compare(time, that.time); -097 if (c != 0) { -098return c; -099 } -100 return Integer.compare(System.identityHashCode(this), System.identityHashCode(that)); -101} -102 -103@Override -104public long getDelay(TimeUnit unit) { -105 return unit.convert(time - System.nanoTime(), TimeUnit.NANOSECONDS); -106} -107 } -108 -109 private static final Task CLOSE = new Task() { -110 }; -111 -112 private final DelayQueueTask tasks = new DelayQueue(); -113 -114 private final AtomicBoolean closed = new AtomicBoolean(false); -115 -116 private ZooKeeper zookeeper; -117 -118 private String getId() { -119return String.format("0x%08x", System.identityHashCode(this)); -120 } -121 -122 public ReadOnlyZKClient(Configuration conf) { -123this.connectString = ZKConfig.getZKQuorumServersString(conf); -124this.sessionTimeoutMs = conf.getInt(ZK_SESSION_TIMEOUT, DEFAULT_ZK_SESSION_TIMEOUT); -125this.maxRetries = conf.getInt(RECOVERY_RETRY, DEFAULT_RECOVERY_RETRY); -126this.retryIntervalMs = -127 conf.getInt(RECOVERY_RETRY_INTERVAL_MILLIS, DEFAULT_RECOVERY_RETRY_INTERVAL_MILLIS); -128this.keepAliveTimeMs = conf.getInt(KEEPALIVE_MILLIS, DEFAULT_KEEPALIVE_MILLIS); -129LOG.info("Start read only zookeeper connection " + getId() + " to " + connectString + -130", session timeout " + sessionTimeoutMs + " ms, retries " + maxRetries + -131", retry interval " + retryIntervalMs + " ms, keep alive " + keepAliveTimeMs + " ms"); -132Thread t = new Thread(this::run, "ReadOnlyZKClient"); -133t.setDaemon(true); -134t.start(); -135 } -136 -137 private abstract class ZKTaskT extends Task { -138 -139protected final String path; -140 -141private final CompletableFutureT future; -142 -143private final String operationType; -144 -145private int retries; -146 -147protected ZKTask(String path, CompletableFutureT future, String operationType) { -148 this.path = path; -149 this.future = future; -150 this.operationType = operationType; -151} -152 -153protected final void onComplete(ZooKeeper zk, int rc, T ret, boolean errorIfNoNode) { -154 tasks.add(new
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/tool/LoadIncrementalHFiles.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/tool/LoadIncrementalHFiles.html b/devapidocs/src-html/org/apache/hadoop/hbase/tool/LoadIncrementalHFiles.html index 1f114e0..01e19b2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/tool/LoadIncrementalHFiles.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/tool/LoadIncrementalHFiles.html @@ -25,1239 +25,1263 @@ 017 */ 018package org.apache.hadoop.hbase.tool; 019 -020import static java.lang.String.format; -021 -022import java.io.FileNotFoundException; -023import java.io.IOException; -024import java.io.InterruptedIOException; -025import java.nio.ByteBuffer; -026import java.util.ArrayDeque; -027import java.util.ArrayList; -028import java.util.Arrays; -029import java.util.Collection; -030import java.util.Collections; -031import java.util.Deque; -032import java.util.HashMap; -033import java.util.HashSet; -034import java.util.List; -035import java.util.Map; -036import java.util.Map.Entry; -037import java.util.Optional; -038import java.util.Set; -039import java.util.SortedMap; -040import java.util.TreeMap; -041import java.util.UUID; -042import java.util.concurrent.Callable; -043import java.util.concurrent.ExecutionException; -044import java.util.concurrent.ExecutorService; -045import java.util.concurrent.Future; -046import java.util.concurrent.LinkedBlockingQueue; -047import java.util.concurrent.ThreadPoolExecutor; -048import java.util.concurrent.TimeUnit; -049import java.util.concurrent.atomic.AtomicInteger; -050import java.util.stream.Collectors; -051 -052import org.apache.commons.lang3.mutable.MutableInt; -053import org.apache.hadoop.conf.Configuration; -054import org.apache.hadoop.conf.Configured; -055import org.apache.hadoop.fs.FileStatus; -056import org.apache.hadoop.fs.FileSystem; -057import org.apache.hadoop.fs.Path; -058import org.apache.hadoop.fs.permission.FsPermission; -059import org.apache.hadoop.hbase.HBaseConfiguration; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.TableName; -062import org.apache.hadoop.hbase.TableNotFoundException; -063import org.apache.yetus.audience.InterfaceAudience; -064import org.slf4j.Logger; -065import org.slf4j.LoggerFactory; -066import org.apache.hadoop.hbase.client.Admin; -067import org.apache.hadoop.hbase.client.ClientServiceCallable; -068import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -069import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -070import org.apache.hadoop.hbase.client.Connection; -071import org.apache.hadoop.hbase.client.ConnectionFactory; -072import org.apache.hadoop.hbase.client.RegionLocator; -073import org.apache.hadoop.hbase.client.RpcRetryingCallerFactory; -074import org.apache.hadoop.hbase.client.SecureBulkLoadClient; -075import org.apache.hadoop.hbase.client.Table; -076import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -077import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -078import org.apache.hadoop.hbase.io.HFileLink; -079import org.apache.hadoop.hbase.io.HalfStoreFileReader; -080import org.apache.hadoop.hbase.io.Reference; -081import org.apache.hadoop.hbase.io.compress.Compression.Algorithm; -082import org.apache.hadoop.hbase.io.hfile.CacheConfig; -083import org.apache.hadoop.hbase.io.hfile.HFile; -084import org.apache.hadoop.hbase.io.hfile.HFileContext; -085import org.apache.hadoop.hbase.io.hfile.HFileContextBuilder; -086import org.apache.hadoop.hbase.io.hfile.HFileDataBlockEncoder; -087import org.apache.hadoop.hbase.io.hfile.HFileScanner; -088import org.apache.hadoop.hbase.ipc.RpcControllerFactory; -089import org.apache.hadoop.hbase.regionserver.BloomType; -090import org.apache.hadoop.hbase.regionserver.HStore; -091import org.apache.hadoop.hbase.regionserver.StoreFileInfo; -092import org.apache.hadoop.hbase.regionserver.StoreFileWriter; -093import org.apache.hadoop.hbase.security.UserProvider; -094import org.apache.hadoop.hbase.security.token.FsDelegationToken; -095import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -096import org.apache.hbase.thirdparty.com.google.common.collect.HashMultimap; -097import org.apache.hbase.thirdparty.com.google.common.collect.Multimap; -098import org.apache.hbase.thirdparty.com.google.common.collect.Multimaps; -099import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -100import org.apache.hadoop.hbase.util.Bytes; -101import org.apache.hadoop.hbase.util.FSHDFSUtils; -102import org.apache.hadoop.hbase.util.Pair; -103import org.apache.hadoop.util.Tool; -104import org.apache.hadoop.util.ToolRunner; -105 -106/** -107 * Tool to load the output of HFileOutputFormat into an existing table. -108 */ -109@InterfaceAudience.Public
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/class-use/ZooKeeperConnectionException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ZooKeeperConnectionException.html b/devapidocs/org/apache/hadoop/hbase/class-use/ZooKeeperConnectionException.html index 9ccf159..bcc708f 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/ZooKeeperConnectionException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/ZooKeeperConnectionException.html @@ -227,11 +227,6 @@ -HBaseFsck(org.apache.hadoop.conf.Configurationconf) -Constructor - - - HBaseFsck(org.apache.hadoop.conf.Configurationconf, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorServiceexec) Constructor
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/devapidocs/org/apache/hadoop/hbase/client/ClusterStatusListener.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/ClusterStatusListener.html b/devapidocs/org/apache/hadoop/hbase/client/ClusterStatusListener.html index b31c22c..7ae608c 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/ClusterStatusListener.html +++ b/devapidocs/org/apache/hadoop/hbase/client/ClusterStatusListener.html @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -class ClusterStatusListener +class ClusterStatusListener extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable A class that receives the cluster status, and provide it as a set of service to the client. @@ -243,7 +243,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h void -receive(ClusterStatusncs) +receive(ClusterMetricsncs) Acts upon the reception of a new cluster status. @@ -275,7 +275,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -284,7 +284,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h deadServers -private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName deadServers +private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName deadServers @@ -293,7 +293,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h deadServerHandler -protected finalClusterStatusListener.DeadServerHandler deadServerHandler +protected finalClusterStatusListener.DeadServerHandler deadServerHandler @@ -302,7 +302,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h listener -private finalClusterStatusListener.Listener listener +private finalClusterStatusListener.Listener listener @@ -311,7 +311,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h STATUS_LISTENER_CLASS -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String STATUS_LISTENER_CLASS +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String STATUS_LISTENER_CLASS The implementation class to use to read the status. See Also: @@ -325,7 +325,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h DEFAULT_STATUS_LISTENER_CLASS -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends ClusterStatusListener.Listener DEFAULT_STATUS_LISTENER_CLASS +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends ClusterStatusListener.Listener DEFAULT_STATUS_LISTENER_CLASS @@ -342,7 +342,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h ClusterStatusListener -publicClusterStatusListener(ClusterStatusListener.DeadServerHandlerdsh, +publicClusterStatusListener(ClusterStatusListener.DeadServerHandlerdsh, org.apache.hadoop.conf.Configurationconf, http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends ClusterStatusListener.ListenerlistenerClass) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -360,13 +360,13 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h Method Detail - + receive -publicvoidreceive(ClusterStatusncs) +publicvoidreceive(ClusterMetricsncs) Acts upon the reception of a new cluster status. Parameters: @@ -380,7 +380,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h close -publicvoidclose() +publicvoidclose() Specified by: http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true#close--; title="class or interface in java.io">closein interfacehttp://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable @@ -395,7 +395,7 @@ implements
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html index 5b3b750..a1f3f7e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html @@ -97,3307 +97,3304 @@ 089import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; 090import org.apache.hbase.thirdparty.io.netty.util.Timeout; 091import org.apache.hbase.thirdparty.io.netty.util.TimerTask; -092import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -093import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -094import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse; -099import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse; -107import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest; -108import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse; -109import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest; -110import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse; -111import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest; -112import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse; -113import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -114import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; -115import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -116import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -117import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -118import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -119import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -120import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -133import
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/class-use/ServerLoad.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ServerLoad.html b/devapidocs/org/apache/hadoop/hbase/class-use/ServerLoad.html index 3d20059..0107f4f 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/ServerLoad.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/ServerLoad.html @@ -368,6 +368,6 @@ -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetrics.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetrics.html b/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetrics.html index fbaf512..f62c0d6 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetrics.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetrics.html @@ -309,6 +309,6 @@ -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.ServerMetricsImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.ServerMetricsImpl.html b/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.ServerMetricsImpl.html index d5bdea0..9a759e5 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.ServerMetricsImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.ServerMetricsImpl.html @@ -120,6 +120,6 @@ -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.html b/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.html index cd402bd..f598397 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/ServerMetricsBuilder.html @@ -204,6 +204,6 @@ -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html b/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html index d9d4ce4..99d28eb 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html @@ -6483,6 +6483,6 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/class-use/ServiceNotRunningException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ServiceNotRunningException.html b/devapidocs/org/apache/hadoop/hbase/class-use/ServiceNotRunningException.html index 6b801c8..65a6318 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/ServiceNotRunningException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/ServiceNotRunningException.html @@ -165,6 +165,6 @@ -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/class-use/SharedConnection.html -- diff --git
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html index d405629..3ec93bb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html @@ -371,1638 +371,1646 @@ 363if (params.getWriteBufferSize() == BufferedMutatorParams.UNSET) { 364 params.writeBufferSize(connectionConfig.getWriteBufferSize()); 365} -366if (params.getMaxKeyValueSize() == BufferedMutatorParams.UNSET) { -367 params.maxKeyValueSize(connectionConfig.getMaxKeyValueSize()); -368} -369// Look to see if an alternate BufferedMutation implementation is wanted. -370// Look in params and in config. If null, use default. -371String implementationClassName = params.getImplementationClassName(); -372if (implementationClassName == null) { -373 implementationClassName = this.alternateBufferedMutatorClassName; -374} -375if (implementationClassName == null) { -376 return new BufferedMutatorImpl(this, rpcCallerFactory, rpcControllerFactory, params); -377} -378try { -379 return (BufferedMutator)ReflectionUtils.newInstance(Class.forName(implementationClassName), -380 this, rpcCallerFactory, rpcControllerFactory, params); -381} catch (ClassNotFoundException e) { -382 throw new RuntimeException(e); -383} -384 } -385 -386 @Override -387 public BufferedMutator getBufferedMutator(TableName tableName) { -388return getBufferedMutator(new BufferedMutatorParams(tableName)); -389 } -390 -391 @Override -392 public RegionLocator getRegionLocator(TableName tableName) throws IOException { -393return new HRegionLocator(tableName, this); -394 } -395 -396 @Override -397 public Admin getAdmin() throws IOException { -398return new HBaseAdmin(this); -399 } -400 -401 @Override -402 public MetricsConnection getConnectionMetrics() { -403return this.metrics; -404 } -405 -406 private ExecutorService getBatchPool() { -407if (batchPool == null) { -408 synchronized (this) { -409if (batchPool == null) { -410 int threads = conf.getInt("hbase.hconnection.threads.max", 256); -411 this.batchPool = getThreadPool(threads, threads, "-shared", null); -412 this.cleanupPool = true; -413} -414 } -415} -416return this.batchPool; -417 } -418 -419 private ExecutorService getThreadPool(int maxThreads, int coreThreads, String nameHint, -420 BlockingQueueRunnable passedWorkQueue) { -421// shared HTable thread executor not yet initialized -422if (maxThreads == 0) { -423 maxThreads = Runtime.getRuntime().availableProcessors() * 8; -424} -425if (coreThreads == 0) { -426 coreThreads = Runtime.getRuntime().availableProcessors() * 8; -427} -428long keepAliveTime = conf.getLong("hbase.hconnection.threads.keepalivetime", 60); -429BlockingQueueRunnable workQueue = passedWorkQueue; -430if (workQueue == null) { -431 workQueue = -432new LinkedBlockingQueue(maxThreads * -433 conf.getInt(HConstants.HBASE_CLIENT_MAX_TOTAL_TASKS, -434 HConstants.DEFAULT_HBASE_CLIENT_MAX_TOTAL_TASKS)); -435 coreThreads = maxThreads; -436} -437ThreadPoolExecutor tpe = new ThreadPoolExecutor( -438coreThreads, -439maxThreads, -440keepAliveTime, -441TimeUnit.SECONDS, -442workQueue, -443 Threads.newDaemonThreadFactory(toString() + nameHint)); -444tpe.allowCoreThreadTimeOut(true); -445return tpe; -446 } -447 -448 private ExecutorService getMetaLookupPool() { -449if (this.metaLookupPool == null) { -450 synchronized (this) { -451if (this.metaLookupPool == null) { -452 //Some of the threads would be used for meta replicas -453 //To start with, threads.max.core threads can hit the meta (including replicas). -454 //After that, requests will get queued up in the passed queue, and only after -455 //the queue is full, a new thread will be started -456 int threads = conf.getInt("hbase.hconnection.meta.lookup.threads.max", 128); -457 this.metaLookupPool = getThreadPool( -458 threads, -459 threads, -460 "-metaLookup-shared-", new LinkedBlockingQueue()); -461} -462 } -463} -464return this.metaLookupPool; -465 } -466 -467
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/client/Get.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/Get.html b/devapidocs/org/apache/hadoop/hbase/client/Get.html index 0792912..07a4779 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/Get.html +++ b/devapidocs/org/apache/hadoop/hbase/client/Get.html @@ -131,7 +131,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Public public class Get extends Query -implements Row, http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableRow +implements Row Used to perform Get operations on a single row. To get everything for a row, instantiate a Get object with the row to get. @@ -226,6 +226,13 @@ implements Fields inherited from classorg.apache.hadoop.hbase.client.OperationWithAttributes ID_ATRIBUTE + + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.client.Row +COMPARATOR + @@ -583,7 +590,7 @@ implements LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -592,7 +599,7 @@ implements row -privatebyte[] row +privatebyte[] row @@ -601,7 +608,7 @@ implements maxVersions -privateint maxVersions +privateint maxVersions @@ -610,7 +617,7 @@ implements cacheBlocks -privateboolean cacheBlocks +privateboolean cacheBlocks @@ -619,7 +626,7 @@ implements storeLimit -privateint storeLimit +privateint storeLimit @@ -628,7 +635,7 @@ implements storeOffset -privateint storeOffset +privateint storeOffset @@ -637,7 +644,7 @@ implements tr -privateTimeRange tr +privateTimeRange tr @@ -646,7 +653,7 @@ implements checkExistenceOnly -privateboolean checkExistenceOnly +privateboolean checkExistenceOnly @@ -655,7 +662,7 @@ implements closestRowBefore -privateboolean closestRowBefore +privateboolean closestRowBefore @@ -664,7 +671,7 @@ implements familyMap -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],http://docs.oracle.com/javase/8/docs/api/java/util/NavigableSet.html?is-external=true; title="class or interface in java.util">NavigableSetbyte[] familyMap +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],http://docs.oracle.com/javase/8/docs/api/java/util/NavigableSet.html?is-external=true; title="class or interface in java.util">NavigableSetbyte[] familyMap @@ -681,7 +688,7 @@ implements Get -publicGet(byte[]row) +publicGet(byte[]row) Create a Get operation for the specified row. If no further operations are done, this will get the latest version of @@ -698,7 +705,7 @@ implements Get -publicGet(Getget) +publicGet(Getget) Copy-constructor Parameters: @@ -712,7 +719,7 @@ implements Get -publicGet(byte[]row, +publicGet(byte[]row, introwOffset, introwLength) Create a Get operation for the specified row. @@ -730,7 +737,7 @@ implements Get -publicGet(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBufferrow) +publicGet(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBufferrow) Create a Get operation for the specified row. Parameters: @@ -752,7 +759,7 @@ implements isCheckExistenceOnly -publicbooleanisCheckExistenceOnly() +publicbooleanisCheckExistenceOnly() @@ -761,7 +768,7 @@ implements setCheckExistenceOnly -publicGetsetCheckExistenceOnly(booleancheckExistenceOnly) +publicGetsetCheckExistenceOnly(booleancheckExistenceOnly) @@ -771,7 +778,7 @@ implements isClosestRowBefore http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true; title="class or interface in java.lang">@Deprecated -publicbooleanisClosestRowBefore() +publicbooleanisClosestRowBefore() Deprecated.since 2.0.0 and will be removed in 3.0.0 This will always return the default value which is false as client cannot set the value to this property any more. @@ -784,7 +791,7 @@ publicboolean setClosestRowBefore http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true; title="class or interface in java.lang">@Deprecated -publicGetsetClosestRowBefore(booleanclosestRowBefore) +publicGetsetClosestRowBefore(booleanclosestRowBefore) Deprecated.since 2.0.0 and will be removed in 3.0.0 This is not used any more and does nothing. Use reverse scan instead. @@ -795,7 +802,7 @@ public addFamily -publicGetaddFamily(byte[]family) +publicGetaddFamily(byte[]family) Get all columns from the specified family. Overrides previous calls to addColumn for this family. @@ -813,7
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowDeleteFamilyCell.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowDeleteFamilyCell.html b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowDeleteFamilyCell.html index 8ff25d5..3a071d6 100644 --- a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowDeleteFamilyCell.html +++ b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowDeleteFamilyCell.html @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab"; PrevClass -NextClass +NextClass Frames @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -private static class PrivateCellUtil.FirstOnRowDeleteFamilyCell +private static class PrivateCellUtil.FirstOnRowDeleteFamilyCell extends PrivateCellUtil.EmptyCell @@ -141,7 +141,7 @@ extends Cell -Cell.DataType +Cell.Type @@ -242,9 +242,9 @@ extends getTimestamp() -Cell.DataType +Cell.Type getType() -Returns the type of cell in a human readable format using Cell.DataType +Returns the type of cell in a human readable format using Cell.Type @@ -304,7 +304,7 @@ extends FIXED_OVERHEAD -private static finalint FIXED_OVERHEAD +private static finalint FIXED_OVERHEAD @@ -313,7 +313,7 @@ extends row -private finalbyte[] row +private finalbyte[] row @@ -322,7 +322,7 @@ extends fam -private finalbyte[] fam +private finalbyte[] fam @@ -339,7 +339,7 @@ extends FirstOnRowDeleteFamilyCell -publicFirstOnRowDeleteFamilyCell(byte[]row, +publicFirstOnRowDeleteFamilyCell(byte[]row, byte[]fam) @@ -357,7 +357,7 @@ extends heapSize -publiclongheapSize() +publiclongheapSize() Returns: Approximate 'exclusive deep size' of implementing object. Includes @@ -371,7 +371,7 @@ extends getRowArray -publicbyte[]getRowArray() +publicbyte[]getRowArray() Description copied from interface:Cell Contiguous raw bytes that may start at any index in the containing array. Max length is Short.MAX_VALUE which is 32,767 bytes. @@ -391,7 +391,7 @@ extends getRowLength -publicshortgetRowLength() +publicshortgetRowLength() Specified by: getRowLengthin interfaceCell @@ -408,7 +408,7 @@ extends getFamilyArray -publicbyte[]getFamilyArray() +publicbyte[]getFamilyArray() Description copied from interface:Cell Contiguous bytes composed of legal HDFS filename characters which may start at any index in the containing array. Max length is Byte.MAX_VALUE, which is 127 bytes. @@ -428,7 +428,7 @@ extends getFamilyLength -publicbytegetFamilyLength() +publicbytegetFamilyLength() Specified by: getFamilyLengthin interfaceCell @@ -445,7 +445,7 @@ extends getTimestamp -publiclonggetTimestamp() +publiclonggetTimestamp() Returns: Long value representing time at which this cell was "Put" into the row. Typically @@ -459,7 +459,7 @@ extends getTypeByte -publicbytegetTypeByte() +publicbytegetTypeByte() Returns: The byte representation of the KeyValue.TYPE of this cell: one of Put, Delete, etc @@ -472,9 +472,9 @@ extends getType -publicCell.DataTypegetType() +publicCell.TypegetType() Description copied from interface:ExtendedCell -Returns the type of cell in a human readable format using Cell.DataType +Returns the type of cell in a human readable format using Cell.Type Note : This does not expose the internal types of Cells like KeyValue.Type.Maximum and KeyValue.Type.Minimum @@ -513,7 +513,7 @@ extends PrevClass -NextClass +NextClass Frames http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowByteBufferCell.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowByteBufferCell.html b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowByteBufferCell.html deleted file mode 100644 index 83b13ff..000 --- a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowByteBufferCell.html +++ /dev/null @@ -1,546 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - - - - -PrivateCellUtil.LastOnRowByteBufferCell (Apache HBase 3.0.0-SNAPSHOT API) - - - - - -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; -var altColor = "altColor"; -var rowColor = "rowColor"; -var tableTab = "tableTab"; -var activeTableTab = "activeTableTab"; - - -JavaScript is disabled on your browser. - - - - - -Skip navigation
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/SizeCachedNoTagsKeyValue.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/SizeCachedNoTagsKeyValue.html b/devapidocs/org/apache/hadoop/hbase/SizeCachedNoTagsKeyValue.html index 4400c9e..2b95af0 100644 --- a/devapidocs/org/apache/hadoop/hbase/SizeCachedNoTagsKeyValue.html +++ b/devapidocs/org/apache/hadoop/hbase/SizeCachedNoTagsKeyValue.html @@ -248,7 +248,7 @@ extends Methods inherited from classorg.apache.hadoop.hbase.KeyValue -checkParameters, clone, create, create, createKeyOnly, deepClone, equals, getBuffer, getDelimiter, getDelimiterInReverse, getFamilyArray, getFamilyLength, getFamilyLength, getFamilyOffset, getKey, getKeyDataStructureSize, getKeyOffset, getKeyString, getKeyValueDataStructureSize, getKeyValueDataStructureSize, getKeyValueDataStructureSize, getLength, getOffset, getQualifierArray, getQualifierLength, getQualifierOffset, getRowArray, getRowOffset, getSeque nceId, getTag, getTags, getTagsArray, getTagsOffset, getTimestamp, getTimestamp, getTimestampOffset, getTypeByte, getValueArray, getValueLength, getValueOffset, hashCode, humanReadableTimestamp, isLatestTimestamp, keyToString, keyToString, oswrite, setSequenceId, setTimestamp, setTimestamp, shallowCopy, toString, toStringMap, updateLatestStamp, write, write, writeByteArray +checkParameters, clone, create, create, createKeyOnly, deepClone, equals, getBuffer, getDelimiter, getDelimiterInReverse, getFamilyArray, getFamilyLength, getFamilyLength, getFamilyOffset, getKey, getKeyDataStructureSize, getKeyOffset, getKeyString, getKeyValueDataStructureSize, getKeyValueDataStructureSize, getKeyValueDataStructureSize, getLength, getOffset, getQualifierArray, getQualifierLength, getQualifierOffset, getRowArray, getRowOffset, getSeque nceId, getTagsArray, getTagsOffset, getTimestamp, getTimestamp, getTimestampOffset, getTypeByte, getValueArray, getValueLength, getValueOffset, hashCode, humanReadableTimestamp, isLatestTimestamp, keyToString, keyToString, oswrite, setSequenceId, setTimestamp, setTimestamp, shallowCopy, toString, toStringMap, updateLatestStamp, write, write, writeByteArray @@ -269,7 +269,7 @@ extends Methods inherited from interfaceorg.apache.hadoop.hbase.RawCell -checkForTagsLength, cloneTags +checkForTagsLength, cloneTags, getTag, getTags http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html index 7d49754..d7fec8e 100644 --- a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html @@ -167,10 +167,10 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase org.apache.hadoop.hbase.backup.BackupType -org.apache.hadoop.hbase.backup.BackupInfo.BackupState org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand +org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase +org.apache.hadoop.hbase.backup.BackupInfo.BackupState http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/class-use/RegionMetrics.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/RegionMetrics.html b/devapidocs/org/apache/hadoop/hbase/class-use/RegionMetrics.html index 868db79..e224d94 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/RegionMetrics.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/RegionMetrics.html @@ -235,7 +235,7 @@ -ServerMetricsImpl(ServerNameserverName, +ServerMetricsImpl(ServerNameserverName, longrequestCountPerSecond, longrequestCount, SizeusedHeapSize, @@ -244,7 +244,7 @@ http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListReplicationLoadSourcesources, ReplicationLoadSinksink, http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/devapidocs/org/apache/hadoop/hbase/client/Mutation.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/Mutation.html b/devapidocs/org/apache/hadoop/hbase/client/Mutation.html index a5e9ff7..55b12f7 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/Mutation.html +++ b/devapidocs/org/apache/hadoop/hbase/client/Mutation.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":9,"i2":9,"i3":9,"i4":9,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":42,"i36":10,"i37":10,"i38":10,"i39":10,"i40":9,"i41":9,"i42":9,"i43":10}; +var methods = {"i0":10,"i1":10,"i2":9,"i3":9,"i4":9,"i5":9,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":42,"i37":10,"i38":10,"i39":10,"i40":10,"i41":9,"i42":9,"i43":9,"i44":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab"; PrevClass -NextClass +NextClass Frames @@ -74,7 +74,7 @@ var activeTableTab = "activeTableTab"; Summary: -Nested| +Nested| Field| Constr| Method @@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Public -public abstract class Mutation +public abstract class Mutation extends OperationWithAttributes implements Row, CellScannable, HeapSize @@ -137,6 +137,25 @@ implements + + + + + +Nested Class Summary + +Nested Classes + +Modifier and Type +Class and Description + + +private static class +Mutation.CellWrapper + + + + @@ -241,39 +260,43 @@ implements Method and Description +(package private) Mutation +add(Cellcell) + + CellScanner cellScanner() - + private static http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object cellToStringMap(Cellc) - + (package private) static byte[] checkRow(byte[]row) - + (package private) static byte[] checkRow(byte[]row, intoffset, intlength) - + (package private) static void checkRow(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBufferrow) - + int compareTo(Rowd) - + (package private) KeyValue createPutKeyValue(byte[]family, byte[]qualifier, longts, byte[]value) - + (package private) KeyValue createPutKeyValue(byte[]family, byte[]qualifier, @@ -283,7 +306,7 @@ implements Create a KeyValue with this objects row key and the Put identifier. - + (package private) KeyValue createPutKeyValue(byte[]family, http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBufferqualifier, @@ -291,75 +314,75 @@ implements http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBuffervalue, Tag[]tags) - + protected long extraHeapSize() Subclasses should override this method to add the heap size of their own fields. - + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCell get(byte[]family, byte[]qualifier) Returns a list of all KeyValue objects with matching column family and qualifier. - + byte[] getACL() - + (package private) http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCell getCellList(byte[]family) Creates an empty list if one doesn't exist for the given column family or else it returns the associated list of Cell objects. - + CellVisibility getCellVisibility() - + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/UUID.html?is-external=true; title="class or interface in java.util">UUID getClusterIds() - + Durability getDurability() Get
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html index b9cd85d..313f199 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html @@ -131,8 +131,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput -org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult +org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl.html b/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl.html new file mode 100644 index 000..fe02a73 --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl.html @@ -0,0 +1,620 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl (Apache HBase 3.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.replication +Class ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.replication.ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl + + + + + + + +All Implemented Interfaces: +ReplicationPeerConfigBuilder + + +Enclosing class: +ReplicationPeerConfig + + + +static class ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +implements ReplicationPeerConfigBuilder + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +private long +bandwidth + + +private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +clusterKey + + +private http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +configuration + + +private http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Sethttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html b/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html index e9f0b87..1a9c79a 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html @@ -27,473 +27,374 @@ 019 020package org.apache.hadoop.hbase; 021 -022import java.util.ArrayList; -023import java.util.Arrays; -024import java.util.Collection; -025import java.util.Collections; +022import edu.umd.cs.findbugs.annotations.Nullable; +023import java.util.ArrayList; +024import java.util.Arrays; +025import java.util.Collection; 026import java.util.List; 027import java.util.Map; -028 -029import org.apache.hadoop.hbase.shaded.com.google.common.base.Objects; +028import java.util.stream.Collectors; +029import org.apache.hadoop.hbase.master.RegionState; 030import org.apache.yetus.audience.InterfaceAudience; -031import org.apache.hadoop.hbase.master.RegionState; -032 -033/** -034 * Status information on the HBase cluster. -035 * p -036 * ttClusterStatus/tt provides clients with information such as: -037 * ul -038 * liThe count and names of region servers in the cluster./li -039 * liThe count and names of dead region servers in the cluster./li -040 * liThe name of the active master for the cluster./li -041 * liThe name(s) of the backup master(s) for the cluster, if they exist./li -042 * liThe average cluster load./li -043 * liThe number of regions deployed on the cluster./li -044 * liThe number of requests since last report./li -045 * liDetailed region server loading and resource usage information, -046 * per server and per region./li -047 * liRegions in transition at master/li -048 * liThe unique cluster ID/li -049 * /ul -050 * tt{@link Option}/tt provides a way to get desired ClusterStatus information. -051 * The following codes will get all the cluster information. -052 * pre -053 * {@code -054 * // Original version still works -055 * Admin admin = connection.getAdmin(); -056 * ClusterStatus status = admin.getClusterStatus(); -057 * // or below, a new version which has the same effects -058 * ClusterStatus status = admin.getClusterStatus(EnumSet.allOf(Option.class)); -059 * } -060 * /pre -061 * If information about live servers is the only wanted. -062 * then codes in the following way: -063 * pre -064 * {@code -065 * Admin admin = connection.getAdmin(); -066 * ClusterStatus status = admin.getClusterStatus(EnumSet.of(Option.LIVE_SERVERS)); -067 * } -068 * /pre -069 */ -070@InterfaceAudience.Public -071public class ClusterStatus { -072 -073 // TODO: remove this in 3.0 -074 private static final byte VERSION = 2; -075 -076 private String hbaseVersion; -077 private MapServerName, ServerLoad liveServers; -078 private ListServerName deadServers; -079 private ServerName master; -080 private ListServerName backupMasters; -081 private ListRegionState intransition; -082 private String clusterId; -083 private String[] masterCoprocessors; -084 private Boolean balancerOn; -085 private int masterInfoPort; -086 -087 /** -088 * Use {@link ClusterStatus.Builder} to construct a ClusterStatus instead. -089 * @deprecated As of release 2.0.0, this will be removed in HBase 3.0.0 -090 * (a href="https://issues.apache.org/jira/browse/HBASE-15511"HBASE-15511/a;). -091 */ -092 @Deprecated -093 public ClusterStatus(final String hbaseVersion, final String clusterid, -094 final MapServerName, ServerLoad servers, -095 final CollectionServerName deadServers, -096 final ServerName master, -097 final CollectionServerName backupMasters, -098 final ListRegionState rit, -099 final String[] masterCoprocessors, -100 final Boolean balancerOn, -101 final int masterInfoPort) { -102// TODO: make this constructor private -103this.hbaseVersion = hbaseVersion; -104this.liveServers = servers; -105this.deadServers = new ArrayList(deadServers); -106this.master = master; -107this.backupMasters = new ArrayList(backupMasters); -108this.intransition = rit; -109this.clusterId = clusterid; -110this.masterCoprocessors = masterCoprocessors; -111this.balancerOn = balancerOn; -112this.masterInfoPort = masterInfoPort; -113 } -114 -115 private ClusterStatus(final String hbaseVersion, final String clusterid, -116 final MapServerName, ServerLoad servers, -117 final ListServerName deadServers, -118 final ServerName master, -119 final ListServerName backupMasters, -120 final ListRegionState rit, -121 final String[] masterCoprocessors, -122 final Boolean balancerOn, -123 final int masterInfoPort) { -124this.hbaseVersion = hbaseVersion; -125this.liveServers
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html index ebbc833..36c4c5c 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html @@ -28,299 +28,298 @@ 020import java.io.IOException; 021import java.lang.reflect.Method; 022import java.util.Map; -023 -024import org.apache.commons.logging.Log; -025import org.apache.commons.logging.LogFactory; -026import org.apache.yetus.audience.InterfaceAudience; -027import org.apache.hadoop.conf.Configuration; -028import org.apache.hadoop.hbase.client.Result; -029import org.apache.hadoop.hbase.client.ResultScanner; -030import org.apache.hadoop.hbase.client.Scan; -031import org.apache.hadoop.hbase.client.ScannerCallable; -032import org.apache.hadoop.hbase.client.Table; -033import org.apache.hadoop.hbase.client.metrics.ScanMetrics; -034import org.apache.hadoop.hbase.DoNotRetryIOException; -035import org.apache.hadoop.hbase.io.ImmutableBytesWritable; -036import org.apache.hadoop.hbase.util.Bytes; -037import org.apache.hadoop.mapreduce.Counter; -038import org.apache.hadoop.mapreduce.InputSplit; -039import org.apache.hadoop.mapreduce.TaskAttemptContext; -040import org.apache.hadoop.util.StringUtils; -041 -042import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting; -043 -044/** -045 * Iterate over an HBase table data, return (ImmutableBytesWritable, Result) -046 * pairs. -047 */ -048@InterfaceAudience.Public -049public class TableRecordReaderImpl { -050 public static final String LOG_PER_ROW_COUNT -051 = "hbase.mapreduce.log.scanner.rowcount"; -052 -053 private static final Log LOG = LogFactory.getLog(TableRecordReaderImpl.class); -054 -055 // HBASE_COUNTER_GROUP_NAME is the name of mapreduce counter group for HBase -056 @VisibleForTesting -057 static final String HBASE_COUNTER_GROUP_NAME = "HBase Counters"; -058 private ResultScanner scanner = null; -059 private Scan scan = null; -060 private Scan currentScan = null; -061 private Table htable = null; -062 private byte[] lastSuccessfulRow = null; -063 private ImmutableBytesWritable key = null; -064 private Result value = null; -065 private TaskAttemptContext context = null; -066 private Method getCounter = null; -067 private long numRestarts = 0; -068 private long numStale = 0; -069 private long timestamp; -070 private int rowcount; -071 private boolean logScannerActivity = false; -072 private int logPerRowCount = 100; -073 -074 /** -075 * Restart from survivable exceptions by creating a new scanner. -076 * -077 * @param firstRow The first row to start at. -078 * @throws IOException When restarting fails. -079 */ -080 public void restart(byte[] firstRow) throws IOException { -081currentScan = new Scan(scan); -082currentScan.withStartRow(firstRow); -083 currentScan.setScanMetricsEnabled(true); -084if (this.scanner != null) { -085 if (logScannerActivity) { -086LOG.info("Closing the previously opened scanner object."); -087 } -088 this.scanner.close(); -089} -090this.scanner = this.htable.getScanner(currentScan); -091if (logScannerActivity) { -092 LOG.info("Current scan=" + currentScan.toString()); -093 timestamp = System.currentTimeMillis(); -094 rowcount = 0; -095} -096 } -097 -098 /** -099 * In new mapreduce APIs, TaskAttemptContext has two getCounter methods -100 * Check if getCounter(String, String) method is available. -101 * @return The getCounter method or null if not available. -102 * @throws IOException -103 */ -104 protected static Method retrieveGetCounterWithStringsParams(TaskAttemptContext context) -105 throws IOException { -106Method m = null; -107try { -108 m = context.getClass().getMethod("getCounter", -109new Class [] {String.class, String.class}); -110} catch (SecurityException e) { -111 throw new IOException("Failed test for getCounter", e); -112} catch (NoSuchMethodException e) { -113 // Ignore -114} -115return m; -116 } -117 -118 /** -119 * Sets the HBase table. -120 * -121 * @param htable The {@link org.apache.hadoop.hbase.HTableDescriptor} to scan. -122 */ -123 public void setHTable(Table htable) { -124Configuration conf = htable.getConfiguration(); -125logScannerActivity = conf.getBoolean( -126 ScannerCallable.LOG_SCANNER_ACTIVITY, false); -127logPerRowCount = conf.getInt(LOG_PER_ROW_COUNT, 100); -128this.htable = htable; -129 } -130 -131 /** -132 * Sets the scan defining the
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dad9a249/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSByteBufferCell.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSByteBufferCell.html b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSByteBufferCell.html index d5bd452f..84f9036 100644 --- a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSByteBufferCell.html +++ b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSByteBufferCell.html @@ -137,7 +137,7 @@ var activeTableTab = "activeTableTab"; -private static class PrivateCellUtil.FirstOnRowColTSByteBufferCell +private static class PrivateCellUtil.FirstOnRowColTSByteBufferCell extends PrivateCellUtil.FirstOnRowColByteBufferCell @@ -145,6 +145,21 @@ extends + + + + + +Nested Class Summary + + + + +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.Cell +Cell.DataType + + + @@ -241,14 +256,14 @@ extends PrivateCellUtil.FirstOnRowByteBufferCell -getRowByteBuffer, getRowLength, getRowPosition, getTypeByte +getRowByteBuffer, getRowLength, getRowPosition, getType, getTypeByte Methods inherited from classorg.apache.hadoop.hbase.PrivateCellUtil.EmptyByteBufferCell -getFamilyArray, getFamilyOffset, getQualifierArray, getQualifierOffset, getRowArray, getRowOffset, getSequenceId, getTagsArray, getTagsByteBuffer, getTagsLength, getTagsOffset, getTagsPosition, getValueArray, getValueByteBuffer, getValueLength, getValueOffset , getValuePosition, setSequenceId, setTimestamp, setTimestamp +getFamilyArray, getFamilyOffset, getQualifierArray, getQualifierOffset, getRowArray, getRowOffset, getSequenceId, getTag, getTags, getTagsArray, getTagsByteBuffer, getTagsLength, getTagsOffset, getTagsPosition, getValueArray, getValueByteBuffer, getValueLength, getValueOffset, getValuePosition, setSequenceId, setTimestamp, setTimestamp @@ -269,7 +284,7 @@ extends RawCell -checkForTagsLength, cloneTags, getTag, getTags +checkForTagsLength, cloneTags @@ -291,7 +306,7 @@ extends FIXED_OVERHEAD -private static finalint FIXED_OVERHEAD +private static finalint FIXED_OVERHEAD @@ -300,7 +315,7 @@ extends ts -privatelong ts +privatelong ts @@ -317,7 +332,7 @@ extends FirstOnRowColTSByteBufferCell -publicFirstOnRowColTSByteBufferCell(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBufferrBuffer, +publicFirstOnRowColTSByteBufferCell(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBufferrBuffer, introffset, shortrlength, http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBufferfBuffer, @@ -343,7 +358,7 @@ extends getTimestamp -publiclonggetTimestamp() +publiclonggetTimestamp() Specified by: getTimestampin interfaceCell @@ -361,7 +376,7 @@ extends heapSize -publiclongheapSize() +publiclongheapSize() Specified by: heapSizein interfaceHeapSize http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dad9a249/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSCell.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSCell.html b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSCell.html index 724a078..1bbe396 100644 --- a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSCell.html +++ b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColTSCell.html @@ -132,7 +132,7 @@ var activeTableTab = "activeTableTab"; -private static class PrivateCellUtil.FirstOnRowColTSCell +private static class PrivateCellUtil.FirstOnRowColTSCell extends PrivateCellUtil.FirstOnRowColCell @@ -140,6 +140,21 @@ extends + + + + + +Nested Class Summary + + + + +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.Cell +Cell.DataType + + + @@ -236,14 +251,14 @@ extends PrivateCellUtil.FirstOnRowCell -getRowArray, getRowLength, getRowOffset, getTypeByte +getRowArray, getRowLength, getRowOffset, getType, getTypeByte Methods inherited from classorg.apache.hadoop.hbase.PrivateCellUtil.EmptyCell -getSequenceId, getTagsArray, getTagsLength, getTagsOffset, getValueArray, getValueLength, getValueOffset, setSequenceId, setTimestamp, setTimestamp +getSequenceId, getTag, getTags, getTagsArray, getTagsLength, getTagsOffset, getValueArray, getValueLength, getValueOffset, setSequenceId, setTimestamp, setTimestamp
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/InvalidFamilyOperationException.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/InvalidFamilyOperationException.html b/apidocs/src-html/org/apache/hadoop/hbase/InvalidFamilyOperationException.html index 04c610a..f2d209a 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/InvalidFamilyOperationException.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/InvalidFamilyOperationException.html @@ -34,7 +34,7 @@ 026 */ 027@InterfaceAudience.Public 028public class InvalidFamilyOperationException extends DoNotRetryIOException { -029 private static final long serialVersionUID = 1L 22 - 1L; +029 private static final long serialVersionUID = (1L 22) - 1L; 030 /** default constructor */ 031 public InvalidFamilyOperationException() { 032super(); http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/MasterNotRunningException.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/MasterNotRunningException.html b/apidocs/src-html/org/apache/hadoop/hbase/MasterNotRunningException.html index f5492a6..f29ca61 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/MasterNotRunningException.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/MasterNotRunningException.html @@ -35,7 +35,7 @@ 027 */ 028@InterfaceAudience.Public 029public class MasterNotRunningException extends IOException { -030 private static final long serialVersionUID = 1L 23 - 1L; +030 private static final long serialVersionUID = (1L 23) - 1L; 031 /** default constructor */ 032 public MasterNotRunningException() { 033super(); http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html b/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html index b24a194..8e3eb2f 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html @@ -34,8 +34,8 @@ 026import java.util.TreeMap; 027import java.util.TreeSet; 028 -029import org.apache.yetus.audience.InterfaceAudience; -030import org.apache.hadoop.hbase.util.Bytes; +029import org.apache.hadoop.hbase.util.Bytes; +030import org.apache.yetus.audience.InterfaceAudience; 031 032/** 033 * Namespace POJO class. Used to represent and define namespaces. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html b/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html index b24a194..8e3eb2f 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html @@ -34,8 +34,8 @@ 026import java.util.TreeMap; 027import java.util.TreeSet; 028 -029import org.apache.yetus.audience.InterfaceAudience; -030import org.apache.hadoop.hbase.util.Bytes; +029import org.apache.hadoop.hbase.util.Bytes; +030import org.apache.yetus.audience.InterfaceAudience; 031 032/** 033 * Namespace POJO class. Used to represent and define namespaces. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/NotServingRegionException.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/NotServingRegionException.html b/apidocs/src-html/org/apache/hadoop/hbase/NotServingRegionException.html index b76ac8a..ac2a556 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/NotServingRegionException.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/NotServingRegionException.html @@ -37,7 +37,7 @@ 029 */ 030@InterfaceAudience.Public 031public class NotServingRegionException extends IOException { -032 private static final long serialVersionUID = 1L 17 - 1L; +032 private static final long serialVersionUID = (1L 17) - 1L; 033 034 /** default constructor */ 035 public NotServingRegionException() { http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/ServerName.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/ServerName.html b/apidocs/src-html/org/apache/hadoop/hbase/ServerName.html index 3bef2d8..4f69045 100644 ---
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html index d35d83f..e75437c 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":9,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":9,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":9,"i46":9,"i47":9,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":9,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":9,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":9,"i47":9,"i48":9,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class RSRpcServices +public class RSRpcServices extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements HBaseRPCErrorHandler, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface, PriorityFunction, ConfigurationObserver Implements the regionserver RPC services. @@ -277,54 +277,58 @@ implements requestCount +protected static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +RESERVOIR_ENABLED_KEY + + private int rowSizeWarnThreshold Row size threshold for multi requests above which a warning is logged - + (package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true; title="class or interface in java.util.concurrent.atomic">LongAdder rpcGetRequestCount - + (package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true; title="class or interface in java.util.concurrent.atomic">LongAdder rpcMultiRequestCount - + (package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true; title="class or interface in java.util.concurrent.atomic">LongAdder rpcMutateRequestCount - + (package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true; title="class or interface in java.util.concurrent.atomic">LongAdder rpcScanRequestCount - + (package private) RpcServerInterface rpcServer - + private int rpcTimeout The RPC timeout period (milliseconds) - + private static http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException SCANNER_ALREADY_CLOSED Deprecated. - + private ScannerIdGenerator scannerIdGenerator - + private int scannerLeaseTimeoutPeriod The lease timeout period for client scanners (milliseconds). - + private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,RSRpcServices.RegionScannerHolder scanners @@ -511,6 +515,14 @@ implements createPriority() +protected RpcServerInterface
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/CacheEvictionStatsBuilder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/CacheEvictionStatsBuilder.html b/devapidocs/src-html/org/apache/hadoop/hbase/CacheEvictionStatsBuilder.html index bbff668..372b3d8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/CacheEvictionStatsBuilder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/CacheEvictionStatsBuilder.html @@ -26,36 +26,45 @@ 018 */ 019package org.apache.hadoop.hbase; 020 -021import org.apache.yetus.audience.InterfaceAudience; -022 -023@InterfaceAudience.Private -024public final class CacheEvictionStatsBuilder { -025 long evictedBlocks = 0; -026 long maxCacheSize = 0; -027 -028 CacheEvictionStatsBuilder() { -029 } -030 -031 public CacheEvictionStatsBuilder withEvictedBlocks(long evictedBlocks) { -032this.evictedBlocks = evictedBlocks; -033return this; -034 } -035 -036 public CacheEvictionStatsBuilder withMaxCacheSize(long maxCacheSize) { -037this.maxCacheSize = maxCacheSize; -038return this; -039 } -040 -041 public CacheEvictionStatsBuilder append(CacheEvictionStats stats) { -042this.evictedBlocks += stats.getEvictedBlocks(); -043this.maxCacheSize += stats.getMaxCacheSize(); -044return this; -045 } -046 -047 public CacheEvictionStats build() { -048return new CacheEvictionStats(this); -049 } -050} +021import java.util.HashMap; +022import java.util.Map; +023 +024import org.apache.yetus.audience.InterfaceAudience; +025 +026@InterfaceAudience.Private +027public final class CacheEvictionStatsBuilder { +028 long evictedBlocks = 0; +029 long maxCacheSize = 0; +030 Mapbyte[], Throwable exceptions = new HashMap(); +031 +032 CacheEvictionStatsBuilder() { +033 } +034 +035 public CacheEvictionStatsBuilder withEvictedBlocks(long evictedBlocks) { +036this.evictedBlocks = evictedBlocks; +037return this; +038 } +039 +040 public CacheEvictionStatsBuilder withMaxCacheSize(long maxCacheSize) { +041this.maxCacheSize = maxCacheSize; +042return this; +043 } +044 +045 public void addException(byte[] regionName, Throwable ie){ +046exceptions.put(regionName, ie); +047 } +048 +049 public CacheEvictionStatsBuilder append(CacheEvictionStats stats) { +050this.evictedBlocks += stats.getEvictedBlocks(); +051this.maxCacheSize += stats.getMaxCacheSize(); +052 this.exceptions.putAll(stats.getExceptions()); +053return this; +054 } +055 +056 public CacheEvictionStats build() { +057return new CacheEvictionStats(this); +058 } +059} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/ExtendedCell.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ExtendedCell.html b/devapidocs/src-html/org/apache/hadoop/hbase/ExtendedCell.html index 0b3d6d7..7329354 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ExtendedCell.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ExtendedCell.html @@ -61,82 +61,84 @@ 053// Key 054PrivateCellUtil.writeFlatKey(this, out); 055 -056// Value -057out.write(getValueArray(), getValueOffset(), getValueLength()); -058 -059// Tags length and tags byte array -060if (withTags getTagsLength() 0) { -061 // Tags length -062 out.write((byte)(0xff (getTagsLength() 8))); -063 out.write((byte)(0xff getTagsLength())); -064 -065 // Tags byte array -066 out.write(getTagsArray(), getTagsOffset(), getTagsLength()); -067} -068 -069return getSerializedSize(withTags); -070 } -071 -072 /** -073 * @param withTags Whether to write tags. -074 * @return Bytes count required to serialize this Cell in a {@link KeyValue} format. -075 * br KeyValue format br -076 * codelt;4 bytes keylengthgt; lt;4 bytes valuelengthgt; lt;2 bytes rowlengthgt; -077 * lt;rowgt; lt;1 byte columnfamilylengthgt; lt;columnfamilygt; lt;columnqualifiergt; -078 * lt;8 bytes timestampgt; lt;1 byte keytypegt; lt;valuegt; lt;2 bytes tagslengthgt; -079 * lt;tagsgt;/code -080 */ -081 // TODO remove the boolean param once HBASE-16706 is done. -082 default int getSerializedSize(boolean withTags) { -083return KeyValueUtil.length(getRowLength(), getFamilyLength(), getQualifierLength(), -084getValueLength(), getTagsLength(), withTags); -085 } -086 -087 /** -088 * Write this Cell into the given buf's offset in a {@link KeyValue} format. -089 * @param buf The buffer where to write the Cell. -090 * @param offset The offset within buffer, to write the Cell. -091 */ -092 default void write(ByteBuffer buf, int offset) { -093KeyValueUtil.appendTo(this, buf, offset, true); -094 } -095 -096 /** -097 * Does a deep copy of the
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index 87fef83..cbcd791 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -26,7 +26,7 @@ under the License. 2007 - 2017 The Apache Software Foundation File: 3453, - Errors: 20619, + Errors: 20567, Warnings: 0, Infos: 0 @@ -4848,6 +4848,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.TestAppendTimeRange.java;>org/apache/hadoop/hbase/coprocessor/TestAppendTimeRange.java + + + 0 + + + 0 + + + 0 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.procedure2.TestProcedureReplayOrder.java;>org/apache/hadoop/hbase/procedure2/TestProcedureReplayOrder.java @@ -8343,7 +8357,7 @@ under the License. 0 - 1 + 0 @@ -9519,7 +9533,7 @@ under the License. 0 - 22 + 26 @@ -12333,7 +12347,7 @@ under the License. 0 - 25 + 23 @@ -14027,7 +14041,7 @@ under the License. 0 - 118 + 122 @@ -18031,7 +18045,7 @@ under the License. 0 - 4 + 2 @@ -24219,7 +24233,7 @@ under the License. 0 - 4 + 0 @@ -25605,7 +25619,7 @@ under the License. 0 - 166 + 156 @@ -26683,7 +26697,7 @@ under the License. 0 - 7 + 6 @@ -37617,7 +37631,7 @@ under the License. 0 - 3 + 2 @@ -43348,20 +43362,6 @@ under the License. - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.ArrayUtils.java;>org/apache/hadoop/hbase/util/ArrayUtils.java - - - 0 - - - 0 - - - 1 - - - - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.thrift2.generated.TAppend.java;>org/apache/hadoop/hbase/thrift2/generated/TAppend.java @@ -44225,7 +44225,7 @@ under the License. 0 - 14 + 0 @@ -46955,7 +46955,7 @@ under the License. 0 - 67 + 46 @@ -47291,7 +47291,7 @@ under the License. 0 - 3 + 0 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/coc.html -- diff --git a/coc.html b/coc.html index 41735fa..c4ff8c8 100644 --- a/coc.html +++ b/coc.html @@ -7,7 +7,7 @@ - + Apache HBase Code of Conduct Policy @@ -380,7 +380,7 @@ email to mailto:priv...@hbase.apache.org;>the priv https://www.apache.org/;>The Apache Software Foundation. All rights reserved. -
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/deprecated-list.html -- diff --git a/devapidocs/deprecated-list.html b/devapidocs/deprecated-list.html index c8b1b9c..b4eba88 100644 --- a/devapidocs/deprecated-list.html +++ b/devapidocs/deprecated-list.html @@ -117,16 +117,6 @@ -org.apache.hadoop.hbase.SettableSequenceId -as of 2.0 and will be removed in 3.0. Use ExtendedCell instead - - - -org.apache.hadoop.hbase.SettableTimestamp -as of 2.0 and will be removed in 3.0. Use ExtendedCell instead - - - org.apache.hadoop.hbase.coprocessor.SingletonCoprocessorService Since 2.0. Will be removed in 3.0 @@ -1171,8 +1161,8 @@ -org.apache.hadoop.hbase.wal.WALKey.getLogSeqNum() -Use WALKey.getSequenceId() +org.apache.hadoop.hbase.wal.WALKeyImpl.getLogSeqNum() +Use WALKeyImpl.getSequenceId() @@ -1330,586 +1320,586 @@ -org.apache.hadoop.hbase.CellUtil.getRowByte(Cell, int) +org.apache.hadoop.hbase.ClusterStatus.getRequestsCount() +As of release 2.0.0, this will be removed in HBase 3.0.0 + Use ClusterStatus.getRequestCount() instead. + +org.apache.hadoop.hbase.CellUtil.getRowByte(Cell, int) + + org.apache.hadoop.hbase.client.Table.getRpcTimeout() use Table.getReadRpcTimeout(TimeUnit) or Table.getWriteRpcTimeout(TimeUnit) instead - + org.apache.hadoop.hbase.client.HTable.getRpcTimeout() - + org.apache.hadoop.hbase.rest.client.RemoteHTable.getRpcTimeout() - + org.apache.hadoop.hbase.MetaTableAccessor.getScanForTableName(Connection, TableName) - + org.apache.hadoop.hbase.client.Scan.getScanMetrics() Use ResultScanner.getScanMetrics() instead. And notice that, please do not use this method and ResultScanner.getScanMetrics() together, the metrics will be messed up. - + org.apache.hadoop.hbase.regionserver.StoreFileReader.getScanner(boolean, boolean) Do not write further code which depends on this call. Instead use getStoreFileScanner() which uses the StoreFileScanner class/interface which is the preferred way to scan a store with higher level concepts. - + org.apache.hadoop.hbase.regionserver.StoreFileReader.getScanner(boolean, boolean, boolean) Do not write further code which depends on this call. Instead use getStoreFileScanner() which uses the StoreFileScanner class/interface which is the preferred way to scan a store with higher level concepts. - + org.apache.hadoop.hbase.ServerName.getServerName(String, int, long) Since 2.0. Use ServerName.valueOf(String, int, long) instead. - + org.apache.hadoop.hbase.ServerName.getServerName(String, long) Since 2.0. Use ServerName.valueOf(String, long) instead. - + org.apache.hadoop.hbase.ServerName.getServerNameLessStartCode(String) Since 2.0. Use ServerName.getAddress() - + org.apache.hadoop.hbase.ServerName.getServerStartcodeFromServerName(String) Since 2.0. Use instance of ServerName to pull out start code. - + org.apache.hadoop.hbase.KeyValue.KVComparator.getShortMidpointKey(byte[], byte[]) Since 0.99.2; - + org.apache.hadoop.hbase.HRegionInfo.getShortNameToLog(HRegionInfo...) As of release 2.0.0, this will be removed in HBase 3.0.0 Use RegionInfo.getShortNameToLog(RegionInfo...). - + org.apache.hadoop.hbase.HRegionInfo.getShortNameToLog(ListHRegionInfo) As of release 2.0.0, this will be removed in HBase 3.0.0 Use RegionInfo.getShortNameToLog(List))}. - + org.apache.hadoop.hbase.util.Bytes.getSize() use Bytes.getLength() instead - + org.apache.hadoop.hbase.io.ImmutableBytesWritable.getSize() use ImmutableBytesWritable.getLength() instead - + org.apache.hadoop.hbase.HRegionInfo.getStartKey(byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0 Use RegionInfo.getStartKey(byte[]). - + org.apache.hadoop.hbase.HRegionInfo.getStartKeyForDisplay(HRegionInfo, Configuration) As of release 2.0.0, this will be removed in HBase 3.0.0 Use RegionInfoDisplay#getStartKeyForDisplay(RegionInfo, Configuration) over in hbase-server module. - + org.apache.hadoop.hbase.ServerLoad.getStorefileIndexSizeInMB() As of release 2.0.0, this will be removed in HBase 3.0.0 Use ServerLoad.getStorefileIndexSizeKB() instead. - + org.apache.hadoop.hbase.RegionLoad.getStorefileIndexSizeMB() As of release 2.0.0, this will be removed in HBase 3.0.0 ((https://issues.apache.org/jira/browse/HBASE-3935;>HBASE-3935)). Use RegionLoad.getStorefileIndexSizeKB() instead. - + org.apache.hadoop.hbase.ServerLoad.getStorefileSizeInMB() As of release 2.0.0, this will be removed in HBase 3.0.0 Use ServerLoad.getStorefileSizeMB() instead. - + org.apache.hadoop.hbase.regionserver.HStore.getStoreHomedir(Path, RegionInfo, byte[]) - +
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html index dbe5063..2e0cc38 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html @@ -329,7 +329,8 @@ implements Description copied from interface:RegionObserver Called before the client performs a Get - Call CoprocessorEnvironment#bypass to skip default actions + Call CoprocessorEnvironment#bypass to skip default actions. + If 'bypass' is set, we skip out on calling any subsequent chained coprocessors. Specified by: preGetOpin interfaceRegionObserver http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html index 5872418..9726790 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html @@ -626,7 +626,8 @@ implements Description copied from interface:RegionObserver Called before the client performs a Get - Call CoprocessorEnvironment#bypass to skip default actions + Call CoprocessorEnvironment#bypass to skip default actions. + If 'bypass' is set, we skip out on calling any subsequent chained coprocessors. Specified by: preGetOpin interfaceRegionObserver @@ -662,7 +663,8 @@ implements Description copied from interface:RegionObserver Called before Increment. - Call CoprocessorEnvironment#bypass to skip default actions + Call CoprocessorEnvironment#bypass to skip default actions. + If 'bypass' is set, we skip out on calling any subsequent chained coprocessors. Note: Do not retain references to any Cells in 'increment' beyond the life of this invocation. If need a Cell reference for later use, copy the cell and use that. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html b/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html index 2c57415..e2ea452 100644 --- a/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html @@ -182,14 +182,14 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.filter.FuzzyRowFilter.SatisfiesCode -org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp -org.apache.hadoop.hbase.filter.RegexStringComparator.EngineType -org.apache.hadoop.hbase.filter.FilterList.Operator -org.apache.hadoop.hbase.filter.Filter.ReturnCode org.apache.hadoop.hbase.filter.FilterWrapper.FilterRowRetCode +org.apache.hadoop.hbase.filter.Filter.ReturnCode +org.apache.hadoop.hbase.filter.FilterList.Operator org.apache.hadoop.hbase.filter.FuzzyRowFilter.Order org.apache.hadoop.hbase.filter.CompareFilter.CompareOp +org.apache.hadoop.hbase.filter.FuzzyRowFilter.SatisfiesCode +org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp +org.apache.hadoop.hbase.filter.RegexStringComparator.EngineType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/org/apache/hadoop/hbase/fs/HFileSystem.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/fs/HFileSystem.html b/devapidocs/org/apache/hadoop/hbase/fs/HFileSystem.html index 4e4e827..760b068 100644 --- a/devapidocs/org/apache/hadoop/hbase/fs/HFileSystem.html +++ b/devapidocs/org/apache/hadoop/hbase/fs/HFileSystem.html @@ -474,7 +474,7 @@ extends org.apache.hadoop.fs.FilterFileSystem HFileSystem publicHFileSystem(org.apache.hadoop.fs.FileSystemfs) Wrap a FileSystem
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html index 3edfbef..9707b2c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html @@ -2459,5936 +2459,5935 @@ 2451 } 2452 2453 for (HStore s : storesToFlush) { -2454MemStoreSize flushableSize = s.getFlushableSize(); -2455 totalSizeOfFlushableStores.incMemStoreSize(flushableSize); -2456 storeFlushCtxs.put(s.getColumnFamilyDescriptor().getName(), -2457 s.createFlushContext(flushOpSeqId, tracker)); -2458// for writing stores to WAL -2459 committedFiles.put(s.getColumnFamilyDescriptor().getName(), null); -2460 storeFlushableSize.put(s.getColumnFamilyDescriptor().getName(), flushableSize); -2461 } -2462 -2463 // write the snapshot start to WAL -2464 if (wal != null !writestate.readOnly) { -2465FlushDescriptor desc = ProtobufUtil.toFlushDescriptor(FlushAction.START_FLUSH, -2466getRegionInfo(), flushOpSeqId, committedFiles); -2467// No sync. Sync is below where no updates lock and we do FlushAction.COMMIT_FLUSH -2468WALUtil.writeFlushMarker(wal, this.getReplicationScope(), getRegionInfo(), desc, false, -2469mvcc); -2470 } -2471 -2472 // Prepare flush (take a snapshot) -2473 for (StoreFlushContext flush : storeFlushCtxs.values()) { -2474flush.prepare(); -2475 } -2476} catch (IOException ex) { -2477 doAbortFlushToWAL(wal, flushOpSeqId, committedFiles); -2478 throw ex; -2479} finally { -2480 this.updatesLock.writeLock().unlock(); -2481} -2482String s = "Finished memstore snapshotting " + this + ", syncing WAL and waiting on mvcc, " + -2483"flushsize=" + totalSizeOfFlushableStores; -2484status.setStatus(s); -2485doSyncOfUnflushedWALChanges(wal, getRegionInfo()); -2486return new PrepareFlushResult(storeFlushCtxs, committedFiles, storeFlushableSize, startTime, -2487flushOpSeqId, flushedSeqId, totalSizeOfFlushableStores); -2488 } -2489 -2490 /** -2491 * Utility method broken out of internalPrepareFlushCache so that method is smaller. -2492 */ -2493 private void logFatLineOnFlush(CollectionHStore storesToFlush, long sequenceId) { -2494if (!LOG.isInfoEnabled()) { -2495 return; -2496} -2497// Log a fat line detailing what is being flushed. -2498StringBuilder perCfExtras = null; -2499if (!isAllFamilies(storesToFlush)) { -2500 perCfExtras = new StringBuilder(); -2501 for (HStore store: storesToFlush) { -2502perCfExtras.append("; ").append(store.getColumnFamilyName()); -2503perCfExtras.append("=") -2504 .append(StringUtils.byteDesc(store.getFlushableSize().getDataSize())); -2505 } -2506} -2507LOG.info("Flushing " + + storesToFlush.size() + "/" + stores.size() + -2508" column families, memstore=" + StringUtils.byteDesc(this.memstoreDataSize.get()) + -2509((perCfExtras != null perCfExtras.length() 0)? perCfExtras.toString(): "") + -2510((wal != null) ? "" : "; WAL is null, using passed sequenceid=" + sequenceId)); -2511 } -2512 -2513 private void doAbortFlushToWAL(final WAL wal, final long flushOpSeqId, -2514 final Mapbyte[], ListPath committedFiles) { -2515if (wal == null) return; -2516try { -2517 FlushDescriptor desc = ProtobufUtil.toFlushDescriptor(FlushAction.ABORT_FLUSH, -2518 getRegionInfo(), flushOpSeqId, committedFiles); -2519 WALUtil.writeFlushMarker(wal, this.getReplicationScope(), getRegionInfo(), desc, false, -2520 mvcc); -2521} catch (Throwable t) { -2522 LOG.warn("Received unexpected exception trying to write ABORT_FLUSH marker to WAL:" + -2523 StringUtils.stringifyException(t)); -2524 // ignore this since we will be aborting the RS with DSE. -2525} -2526// we have called wal.startCacheFlush(), now we have to abort it -2527 wal.abortCacheFlush(this.getRegionInfo().getEncodedNameAsBytes()); -2528 } -2529 -2530 /** -2531 * Sync unflushed WAL changes. See HBASE-8208 for details -2532 */ -2533 private static void doSyncOfUnflushedWALChanges(final WAL wal, final RegionInfo hri) -2534 throws IOException { -2535if (wal == null) { -2536 return; -2537} -2538try { -2539 wal.sync(); // ensure that flush marker is sync'ed -2540} catch (IOException ioe) { -2541
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html index bd4be7d..3cbdb37 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html @@ -123,23 +123,25 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public abstract class RegionTransitionProcedure +public abstract class RegionTransitionProcedure extends ProcedureMasterProcedureEnv implements TableProcedureInterface, RemoteProcedureDispatcher.RemoteProcedureMasterProcedureEnv,ServerName Base class for the Assign and Unassign Procedure. - There can only be one RegionTransitionProcedure per region running at a time - since each procedure takes a lock on the region (see MasterProcedureScheduler). + + Locking: + Takes exclusive lock on the region being assigned/unassigned. Thus, there can only be one + RegionTransitionProcedure per region running at a time (see MasterProcedureScheduler). This procedure is asynchronous and responds to external events. The AssignmentManager will notify this procedure when the RS completes the operation and reports the transitioned state - (see the Assign and Unassign class for more detail). + (see the Assign and Unassign class for more detail). Procedures move from the REGION_TRANSITION_QUEUE state when they are first submitted, to the REGION_TRANSITION_DISPATCH state when the request to remote server is sent and the Procedure is suspended waiting on external event to be woken again. Once the external event is triggered, Procedure - moves to the REGION_TRANSITION_FINISH state. + moves to the REGION_TRANSITION_FINISH state. NOTE: AssignProcedure and UnassignProcedure should not be thought of as being asymmetric, at least currently. @@ -152,12 +154,13 @@ implements AssignProcedure.forceNewPlan. When - the number of attempts reach hreshold configuration 'hbase.assignment.maximum.attempts', + the number of attempts reaches threshold configuration 'hbase.assignment.maximum.attempts', the procedure is aborted. For UnassignProcedure, similar re-attempts are intentionally not implemented. It is a 'one shot' procedure. See its class doc for how it handles failure. + TODO: Considering it is a priority doing all we can to get make a region available as soon as possible, re-attempting with any target makes sense if specified target fails in case of @@ -300,7 +303,7 @@ implements RegionStates.RegionStateNoderegionNode) -RegionInfo +protected RegionInfo getRegionInfo() @@ -318,7 +321,7 @@ implements getTableName() -protected org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RegionTransitionState +(package private) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RegionTransitionState getTransitionState() @@ -405,7 +408,7 @@ implements setRegionInfo(RegionInforegionInfo) -protected void +(package private) void setTransitionState(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RegionTransitionStatestate) @@ -482,7 +485,7 @@ implements LOG -private static finalorg.apache.commons.logging.Log LOG +private static finalorg.apache.commons.logging.Log LOG @@ -491,7 +494,7 @@ implements aborted -protected finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicBoolean.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicBoolean aborted +protected finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicBoolean.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicBoolean aborted @@ -500,7 +503,7 @@ implements transitionState -privateorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RegionTransitionState transitionState +privateorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RegionTransitionState transitionState @@ -509,7 +512,7 @@ implements regionInfo -privateRegionInfo regionInfo +privateRegionInfo regionInfo @@ -518,7 +521,7 @@ implements lock -private volatileboolean lock +private volatileboolean lock @@ -535,7 +538,7 @@ implements RegionTransitionProcedure -publicRegionTransitionProcedure() +publicRegionTransitionProcedure() @@ -561,7 +564,7 @@ implements getRegionInfo -publicRegionInfogetRegionInfo() +protectedRegionInfogetRegionInfo() @@ -630,7 +633,7 @@ implements setTransitionState
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html index 2f81057..d5bcd27 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html +++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":9,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":42,"i27":42,"i28":42,"i29":42,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":42,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":9,"i80":10,"i81":10,"i82":9,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":41,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":42,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109 ":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":9,"i116":10,"i117":10,"i118":10,"i119":42,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":42,"i172":10,"i173":10,"i174":10,"i175":10,"i176":10,"i177":10,"i178":10,"i179":10,"i180":10,"i181":10,"i182":10,"i183":10,"i184":10,"i185":10,"i186":10,"i187":10,"i188":10,"i189":10,"i190":10,"i191":10,"i192":10,"i193":10,"i194":10,"i195":10,"i196":10,"i197":42,"i198":10,"i199":10,"i200":10,"i201":10,"i202":10,"i203":10,"i204":10,"i205":10,"i206":10,"i207":10,"i208":10,"i2 09":10,"i210":10,"i211":10,"i212":10,"i213":10,"i214":10,"i215":10,"i216":10,"i217":10,"i218":10,"i219":10,"i220":10,"i221":10,"i222":10,"i223":10,"i224":10,"i225":10,"i226":10,"i227":10,"i228":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":9,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":42,"i27":42,"i28":42,"i29":42,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":42,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":9,"i80":10,"i81":10,"i82":9,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":41,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":42,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109 ":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":9,"i115":10,"i116":10,"i117":10,"i118":42,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":42,"i171":10,"i172":10,"i173":10,"i174":10,"i175":10,"i176":10,"i177":10,"i178":10,"i179":10,"i180":10,"i181":10,"i182":10,"i183":10,"i184":10,"i185":10,"i186":10,"i187":10,"i188":10,"i189":10,"i190":10,"i191":10,"i192":10,"i193":10,"i194":10,"i195":10,"i196":42,"i197":10,"i198":10,"i199":10,"i200":10,"i201":10,"i202":10,"i203":10,"i204":10,"i205":10,"i206":10,"i207":10,"i208":10,"i2 09":10,"i210":10,"i211":10,"i212":10,"i213":10,"i214":10,"i215":10,"i216":10,"i217":10,"i218":10,"i219":10,"i220":10,"i221":10,"i222":10,"i223":10,"i224":10,"i225":10,"i226":10,"i227":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9722a17/devapidocs/org/apache/hadoop/hbase/security/access/CoprocessorWhitelistMasterObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/access/CoprocessorWhitelistMasterObserver.html b/devapidocs/org/apache/hadoop/hbase/security/access/CoprocessorWhitelistMasterObserver.html index 92ebab8..30936b3 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/access/CoprocessorWhitelistMasterObserver.html +++ b/devapidocs/org/apache/hadoop/hbase/security/access/CoprocessorWhitelistMasterObserver.html @@ -255,7 +255,7 @@ implements MasterObserver -postAbortProcedure, postAddReplicationPeer, postAddRSGroup, postAssign, postBalance, postBalanceRSGroup, postBalanceSwitch, postClearDeadServers, postCloneSnapshot, postCompletedCreateTableAction, postCompletedDeleteTableAction, postCompletedDisableTableAction, postCompletedEnableTableAction, postCompletedMergeRegionsAction, postCompletedModifyTableAction, postCompletedSplitRegionAction, postCompletedTruncateTab leAction, postCreateNamespace, postCreateTable, postDecommissionRegionServers, postDeleteNamespace, postDeleteSnapshot, postDeleteTable, postDisableReplicationPeer, postDisableTable, postEnableReplicationPe er, postEnableTable, postGetClusterStatus, postGetLocks, postGetNamespaceDescriptor, postGetProcedures, postGetReplicationPeerConfig, postGetTableDescriptors, postGetTableNames, postListDecommissionedRegionServers, postListNamespaceDescriptors, postListReplicationPeers, postListSnapshot, postLockHeartbeat, postMergeRegions, postMergeRegionsCommitAction, postModifyNamespace, postModifyTable, postMove , postMoveServers, postMoveServersAndTables, postMoveTables, postRecommissionRegionServer, postRegionOffline, postRemoveReplicationPeer, postRemoveRSGroup, postRequestLock, postRestoreSnapshot, postRollBackMergeRegionsAction, postRollBackSplitRegionAction, postSetNamespaceQuota, postSetSplitOrMergeEnabled, postSetTableQuota, postSetUserQuota, postSetUserQuota, postSetUserQuota, postSnapshot, postStartMaster, postTableFlush, postTruncateTable, postUnassign, postUpdateReplicationPeerConfig, preAbortProcedure, preAddReplicationPeer, preAddRSGroup, preAssign, preBalance, preBalanceRSGroup, preBalanceSwitch, preClearDeadServers, preCloneSnapshot, preCreateNamespace, preCreateTableAction, preDec ommissionRegionServers, preDeleteNamespace, preDeleteSnapshot, preDeleteTable, preDeleteTableAction, preDisableReplicationPeer, preDisableTable, preDisableTableAction, preEnableReplicationPeer, preEnableTable, preEnableTableAction, preGetClusterStatus, preGetLocks, preGetNamespaceDescriptor, preGetProcedures, preGetReplicationPeerConfig, preGetTableDescriptors, preGetTableNames, preListDecommissionedRegionServers, preListNamespaceDescriptors, preListReplicationPeers, preListSnapshot, preLockHeartbeat, preMasterInitialization, preMergeRegions, preMergeRegionsAction, preMergeRegionsCommitAction, preModifyNamespace, preModifyTableAction, preMove, preMoveServers, preMoveServersAndTables, preMoveTables, preRecommissionRegionServer, preRegionOffline, preRemoveReplicationPeer, preRemoveRSGroup, preRequestLock, preRestoreSnapshot, preSetNamespaceQuota, preSetSplitOrMergeEnabled, preSetTableQuota, preSetUserQuota, preSetUserQuota, preSetUserQuota, preShutdown, preSnapshot, preSplitRegion, preSplitRegionAction, preSplitRegionAfterMETAAction, preSplitRegionBeforeMETAAction, preStopMaster, preTableFlush, preTruncateTable, preTruncateTableAction, preUnassign, preUpdateReplicationPeerConfig +postAbortProcedure, postAddReplicationPeer, postAddRSGroup, postAssign, postBalance, postBalanceRSGroup, postBalanceSwitch, postClearDeadServers, postCloneSnapshot, postCompletedCreateTableAction, postCompletedDeleteTableAction, postCompletedDisableTableAction, postCompletedEnableTableAction, postCompletedMergeRegionsAction, postCompletedModifyTableAction, postCompletedSplitRegionAction, postCompletedTruncateTableAction, postCreateNamespace, postCreateTable, postDecommissionRegionServers, postDeleteNamespace, postDeleteSnapshot, postDeleteTable, postDisableReplicationPeer, postDisableTable, postEnableReplicationPeer, postEnableTable, postGetClusterStatus, postGetLocks, postGetNamespaceDescriptor, postGetProcedures, postGetReplicationPeerConfig, postGetTableDescriptors, postGetTableNames, postListDecommissionedRegionServers, postListNamespaceDescriptors, postListReplicationPeers, postListSnapshot, postLockHeartbeat, postMergeRegions, postMergeRegionsCommitAction, postModifyNamespace, postModifyTable, postMove, postMoveServers, postMoveServersAndTables, postMoveTables, postRecommissionRegionServer, postRegionOffline, postRemoveReplicationPeer, postRemoveRSGroup, postRequestLock,