[07/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.

2019-01-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftClientBuilder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftClientBuilder.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftClientBuilder.html
new file mode 100644
index 000..875bc4b
--- /dev/null
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftClientBuilder.html
@@ -0,0 +1,109 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/**
+002 *
+003 * Licensed to the Apache Software 
Foundation (ASF) under one
+004 * or more contributor license 
agreements.  See the NOTICE file
+005 * distributed with this work for 
additional information
+006 * regarding copyright ownership.  The 
ASF licenses this file
+007 * to you under the Apache License, 
Version 2.0 (the
+008 * "License"); you may not use this file 
except in compliance
+009 * with the License.  You may obtain a 
copy of the License at
+010 *
+011 * 
http://www.apache.org/licenses/LICENSE-2.0
+012 *
+013 * Unless required by applicable law or 
agreed to in writing, software
+014 * distributed under the License is 
distributed on an "AS IS" BASIS,
+015 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+016 * See the License for the specific 
language governing permissions and
+017 * limitations under the License.
+018 */
+019package 
org.apache.hadoop.hbase.thrift2.client;
+020
+021import java.io.IOException;
+022
+023import 
org.apache.hadoop.hbase.thrift2.generated.THBaseService;
+024import 
org.apache.hadoop.hbase.util.Pair;
+025import 
org.apache.thrift.transport.TTransport;
+026import 
org.apache.yetus.audience.InterfaceAudience;
+027
+028@InterfaceAudience.Private
+029public abstract class ThriftClientBuilder 
{
+030  protected ThriftConnection 
connection;
+031
+032  public 
ThriftClientBuilder(ThriftConnection connection) {
+033this.connection = connection;
+034  }
+035
+036  public abstract 
PairTHBaseService.Client, TTransport getClient() throws IOException;
+037}
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftConnection.DefaultThriftClientBuilder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftConnection.DefaultThriftClientBuilder.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftConnection.DefaultThriftClientBuilder.html
new file mode 100644
index 000..fee41a8
--- /dev/null
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftConnection.DefaultThriftClientBuilder.html
@@ -0,0 +1,394 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/**
+002 *
+003 * Licensed to the Apache Software 
Foundation (ASF) under one
+004 * or more contributor license 
agreements.  See the NOTICE file
+005 * distributed with this work for 
additional information
+006 * regarding copyright ownership.  The 
ASF licenses this file
+007 * to you under the Apache License, 
Version 2.0 (the
+008 * "License"); you may not use this file 
except in compliance
+009 * with the License.  You may obtain a 
copy of the License at
+010 *
+011 * 
http://www.apache.org/licenses/LICENSE-2.0
+012 *
+013 * Unless required by applicable law or 
agreed to in writing, software
+014 * distributed under the License is 
distributed on an "AS IS" BASIS,
+015 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+016 * See the License for the specific 
language governing permissions and
+017 * limitations under the License.
+018 */
+019package 
org.apache.hadoop.hbase.thrift2.client;
+020
+021import static 
org.apache.hadoop.hbase.ipc.RpcClient.DEFAULT_SOCKET_TIMEOUT_CONNECT;
+022import static 
org.apache.hadoop.hbase.ipc.RpcClient.SOCKET_TIMEOUT_CONNECT;
+023
+024import java.io.IOException;
+025import java.lang.reflect.Constructor;
+026import java.util.HashMap;
+027import java.util.Map;
+028import 
java.util.concurrent.ExecutorService;
+029
+030import 
org.apache.commons.lang3.NotImplementedException;
+031import 
org.apache.hadoop.conf.Configuration;
+032import 
org.apache.hadoop.hbase.HConstants;
+033import 
org.apache.hadoop.hbase.TableName;
+034import 
org.apache.hadoop.hbase.client.Admin;
+035import 
org.apache.hadoop.hbase.client.BufferedMutator;
+036import 
org.apache.hadoop.hbase.client.BufferedMutatorParams;
+037import 
org.apache.hadoop.hbase.client.Connection;
+038import 
org.apache.hadoop.hbase.client.RegionLocator;
+039import 
org.apache.hadoop.hbase.client.Table;
+040import 
org.apache.hadoop.hbase.client.TableBuilder;
+041import 

[07/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.

2019-01-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
index 65d5f76..c9b8a49 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
@@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.LimitedPrivate(value="Tools")
-public class HRegionServer
+public class HRegionServer
 extends HasThread
 implements RegionServerServices, LastSequenceId, 
ConfigurationObserver
 HRegionServer makes a set of HRegions available to clients. 
It checks in with
@@ -1554,7 +1554,7 @@ implements 
 
 INIT_PAUSE_TIME_MS
-private static finalint INIT_PAUSE_TIME_MS
+private static finalint INIT_PAUSE_TIME_MS
 
 See Also:
 Constant
 Field Values
@@ -1567,7 +1567,7 @@ implements 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
@@ -1576,7 +1576,7 @@ implements 
 
 TEST_SKIP_REPORTING_TRANSITION
-public staticboolean TEST_SKIP_REPORTING_TRANSITION
+public staticboolean TEST_SKIP_REPORTING_TRANSITION
 For testing only!  Set to true to skip notifying region 
assignment to master .
 
 
@@ -1586,7 +1586,7 @@ implements 
 
 regionsInTransitionInRS
-protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
 title="class or interface in 
java.util.concurrent">ConcurrentMapbyte[],https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean regionsInTransitionInRS
+protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
 title="class or interface in 
java.util.concurrent">ConcurrentMapbyte[],https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean regionsInTransitionInRS
 
 
 
@@ -1595,7 +1595,7 @@ implements 
 
 cacheFlusher
-protectedMemStoreFlusher cacheFlusher
+protectedMemStoreFlusher cacheFlusher
 
 
 
@@ -1604,7 +1604,7 @@ implements 
 
 hMemManager
-protectedHeapMemoryManager hMemManager
+protectedHeapMemoryManager hMemManager
 
 
 
@@ -1613,7 +1613,7 @@ implements 
 
 clusterConnection
-protectedClusterConnection clusterConnection
+protectedClusterConnection clusterConnection
 Cluster connection to be shared by services.
  Initialized at server startup and closed when server shuts down.
  Clients must never close it explicitly.
@@ -1628,7 +1628,7 @@ implements 
 
 tableDescriptors
-protectedTableDescriptors tableDescriptors
+protectedTableDescriptors tableDescriptors
 Go here to get table descriptors.
 
 
@@ -1638,7 +1638,7 @@ implements 
 
 replicationSourceHandler
-protectedReplicationSourceService replicationSourceHandler
+protectedReplicationSourceService replicationSourceHandler
 
 
 
@@ -1647,7 +1647,7 @@ implements 
 
 replicationSinkHandler
-protectedReplicationSinkService replicationSinkHandler
+protectedReplicationSinkService replicationSinkHandler
 
 
 
@@ -1656,7 +1656,7 @@ implements 
 
 compactSplitThread
-publicCompactSplit compactSplitThread
+publicCompactSplit compactSplitThread
 
 
 
@@ -1665,7 +1665,7 @@ implements 
 
 onlineRegions
-protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,HRegion onlineRegions
+protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,HRegion onlineRegions
 Map of regions currently being served by this region 
server. Key is the
  encoded region name.  All access should be synchronized.
 
@@ -1676,7 +1676,7 @@ implements 
 
 regionFavoredNodesMap
-protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/net/InetSocketAddress.html?is-external=true;
 title="class or interface in java.net">InetSocketAddress[] regionFavoredNodesMap
+protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 

[07/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.

2019-01-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html
index 736388b..197b99d 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html
@@ -26,3624 +26,3599 @@
 018package org.apache.hadoop.hbase.client;
 019
 020import static 
org.apache.hadoop.hbase.TableName.META_TABLE_NAME;
-021
-022import com.google.protobuf.Message;
-023import com.google.protobuf.RpcChannel;
-024import java.io.IOException;
-025import java.util.ArrayList;
-026import java.util.Arrays;
-027import java.util.Collections;
-028import java.util.EnumSet;
-029import java.util.HashMap;
-030import java.util.List;
-031import java.util.Map;
-032import java.util.Optional;
-033import java.util.Set;
-034import 
java.util.concurrent.CompletableFuture;
-035import 
java.util.concurrent.ConcurrentHashMap;
-036import java.util.concurrent.TimeUnit;
-037import 
java.util.concurrent.atomic.AtomicReference;
-038import java.util.function.BiConsumer;
-039import java.util.function.Function;
-040import java.util.function.Supplier;
-041import java.util.regex.Pattern;
-042import java.util.stream.Collectors;
-043import java.util.stream.Stream;
-044import org.apache.commons.io.IOUtils;
-045import 
org.apache.hadoop.conf.Configuration;
-046import 
org.apache.hadoop.hbase.AsyncMetaTableAccessor;
-047import 
org.apache.hadoop.hbase.CacheEvictionStats;
-048import 
org.apache.hadoop.hbase.CacheEvictionStatsAggregator;
-049import 
org.apache.hadoop.hbase.ClusterMetrics;
-050import 
org.apache.hadoop.hbase.ClusterMetrics.Option;
-051import 
org.apache.hadoop.hbase.ClusterMetricsBuilder;
-052import 
org.apache.hadoop.hbase.HConstants;
-053import 
org.apache.hadoop.hbase.HRegionLocation;
-054import 
org.apache.hadoop.hbase.MetaTableAccessor;
-055import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
-056import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-057import 
org.apache.hadoop.hbase.RegionLocations;
-058import 
org.apache.hadoop.hbase.RegionMetrics;
-059import 
org.apache.hadoop.hbase.RegionMetricsBuilder;
-060import 
org.apache.hadoop.hbase.ServerName;
-061import 
org.apache.hadoop.hbase.TableExistsException;
-062import 
org.apache.hadoop.hbase.TableName;
-063import 
org.apache.hadoop.hbase.TableNotDisabledException;
-064import 
org.apache.hadoop.hbase.TableNotEnabledException;
-065import 
org.apache.hadoop.hbase.TableNotFoundException;
-066import 
org.apache.hadoop.hbase.UnknownRegionException;
-067import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
-068import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
-069import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.ServerRequestCallerBuilder;
-070import 
org.apache.hadoop.hbase.client.Scan.ReadType;
-071import 
org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil;
-072import 
org.apache.hadoop.hbase.client.replication.TableCFs;
-073import 
org.apache.hadoop.hbase.client.security.SecurityCapability;
-074import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-075import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-076import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-077import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-078import 
org.apache.hadoop.hbase.quotas.QuotaTableUtil;
-079import 
org.apache.hadoop.hbase.replication.ReplicationException;
-080import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-081import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-082import 
org.apache.hadoop.hbase.replication.SyncReplicationState;
-083import 
org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
-084import 
org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
-085import 
org.apache.hadoop.hbase.snapshot.SnapshotCreationException;
-086import 
org.apache.hadoop.hbase.util.Bytes;
-087import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-088import 
org.apache.hadoop.hbase.util.ForeignExceptionUtil;
-089import 
org.apache.yetus.audience.InterfaceAudience;
-090import org.slf4j.Logger;
-091import org.slf4j.LoggerFactory;
-092
-093import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-094import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
-095import 
org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback;
-096import 

[07/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.

2019-01-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/pseudo-distributed.html
--
diff --git a/pseudo-distributed.html b/pseudo-distributed.html
index 0b77d22..c338bde 100644
--- a/pseudo-distributed.html
+++ b/pseudo-distributed.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase   
 Running Apache HBase (TM) in pseudo-distributed mode
@@ -318,7 +318,7 @@ under the License. -->
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2019-01-01
+  Last Published: 
2019-01-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/replication.html
--
diff --git a/replication.html b/replication.html
index e594006..85208d5 100644
--- a/replication.html
+++ b/replication.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  
   Apache HBase (TM) Replication
@@ -313,7 +313,7 @@ under the License. -->
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2019-01-01
+  Last Published: 
2019-01-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/resources.html
--
diff --git a/resources.html b/resources.html
index c4cf437..4cd587f 100644
--- a/resources.html
+++ b/resources.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Other Apache HBase (TM) Resources
 
@@ -341,7 +341,7 @@ under the License. -->
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2019-01-01
+  Last Published: 
2019-01-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/source-repository.html
--
diff --git a/source-repository.html b/source-repository.html
index 00f6a3a..39adeb7 100644
--- a/source-repository.html
+++ b/source-repository.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Source Code Management
 
@@ -309,7 +309,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2019-01-01
+  Last Published: 
2019-01-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/sponsors.html
--
diff --git a/sponsors.html b/sponsors.html
index 692ec94..0129e39 100644
--- a/sponsors.html
+++ b/sponsors.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Apache HBase™ Sponsors
 
@@ -343,7 +343,7 @@ under the License. -->
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2019-01-01
+  Last Published: 
2019-01-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/supportingprojects.html
--
diff --git a/supportingprojects.html b/supportingprojects.html
index 06b9dea..1192b46 100644
--- a/supportingprojects.html
+++ b/supportingprojects.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Supporting Projects
 
@@ -530,7 +530,7 @@ under the License. -->
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2019-01-01
+  Last Published: 
2019-01-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/team-list.html
--
diff --git a/team-list.html b/team-list.html
index cfe8d06..4f4df99 100644
--- a/team-list.html
+++ b/team-list.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Project Team
 
@@ -776,7 +776,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2019-01-01
+  Last Published: 
2019-01-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/testdevapidocs/allclasses-frame.html

[07/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.

2019-01-01 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.html 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.html
index 4a55269..284ab48 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.html
@@ -597,6 +597,6 @@ implements org.apache.hadoop.conf.Configurable
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReader.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReader.html 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReader.html
index c35890c..9a8b447 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReader.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReader.html
@@ -533,6 +533,6 @@ extends org.apache.hadoop.mapreduce.RecordReaderCopyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html
index 3b1c29c..33a2fdc 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.html
@@ -792,6 +792,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/mapreduce/TableReducer.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableReducer.html 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableReducer.html
index a58a429..76c1df3 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableReducer.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableReducer.html
@@ -285,6 +285,6 @@ extends 
org.apache.hadoop.mapreduce.ReducerKEYIN,VALUEIN,KEYOUT,Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionRecordReader.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionRecordReader.html
 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionRecordReader.html
index c79a362..f56f050 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionRecordReader.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionRecordReader.html
@@ -465,6 +465,6 @@ extends org.apache.hadoop.mapreduce.RecordReaderCopyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionSplit.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionSplit.html
 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionSplit.html
index fd03a3e..fdf3be3 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionSplit.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.TableSnapshotRegionSplit.html
@@ -474,6 +474,6 @@ 

[07/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.

2018-12-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
index 0f5a095..50bf692 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
@@ -78,8712 +78,8714 @@
 070import 
java.util.concurrent.locks.ReadWriteLock;
 071import 
java.util.concurrent.locks.ReentrantReadWriteLock;
 072import java.util.function.Function;
-073import 
org.apache.hadoop.conf.Configuration;
-074import org.apache.hadoop.fs.FileStatus;
-075import org.apache.hadoop.fs.FileSystem;
-076import 
org.apache.hadoop.fs.LocatedFileStatus;
-077import org.apache.hadoop.fs.Path;
-078import org.apache.hadoop.hbase.Cell;
-079import 
org.apache.hadoop.hbase.CellBuilderType;
-080import 
org.apache.hadoop.hbase.CellComparator;
-081import 
org.apache.hadoop.hbase.CellComparatorImpl;
-082import 
org.apache.hadoop.hbase.CellScanner;
-083import 
org.apache.hadoop.hbase.CellUtil;
-084import 
org.apache.hadoop.hbase.CompareOperator;
-085import 
org.apache.hadoop.hbase.CompoundConfiguration;
-086import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-087import 
org.apache.hadoop.hbase.DroppedSnapshotException;
-088import 
org.apache.hadoop.hbase.ExtendedCellBuilderFactory;
-089import 
org.apache.hadoop.hbase.HConstants;
-090import 
org.apache.hadoop.hbase.HConstants.OperationStatusCode;
-091import 
org.apache.hadoop.hbase.HDFSBlocksDistribution;
-092import 
org.apache.hadoop.hbase.KeyValue;
-093import 
org.apache.hadoop.hbase.KeyValueUtil;
-094import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-095import 
org.apache.hadoop.hbase.NotServingRegionException;
-096import 
org.apache.hadoop.hbase.PrivateCellUtil;
-097import 
org.apache.hadoop.hbase.RegionTooBusyException;
-098import org.apache.hadoop.hbase.Tag;
-099import org.apache.hadoop.hbase.TagUtil;
-100import 
org.apache.hadoop.hbase.UnknownScannerException;
-101import 
org.apache.hadoop.hbase.client.Append;
-102import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
-103import 
org.apache.hadoop.hbase.client.CompactionState;
-104import 
org.apache.hadoop.hbase.client.Delete;
-105import 
org.apache.hadoop.hbase.client.Durability;
-106import 
org.apache.hadoop.hbase.client.Get;
-107import 
org.apache.hadoop.hbase.client.Increment;
-108import 
org.apache.hadoop.hbase.client.IsolationLevel;
-109import 
org.apache.hadoop.hbase.client.Mutation;
-110import 
org.apache.hadoop.hbase.client.PackagePrivateFieldAccessor;
-111import 
org.apache.hadoop.hbase.client.Put;
-112import 
org.apache.hadoop.hbase.client.RegionInfo;
-113import 
org.apache.hadoop.hbase.client.RegionInfoBuilder;
-114import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-115import 
org.apache.hadoop.hbase.client.Result;
-116import 
org.apache.hadoop.hbase.client.RowMutations;
-117import 
org.apache.hadoop.hbase.client.Scan;
-118import 
org.apache.hadoop.hbase.client.TableDescriptor;
-119import 
org.apache.hadoop.hbase.client.TableDescriptorBuilder;
-120import 
org.apache.hadoop.hbase.conf.ConfigurationManager;
-121import 
org.apache.hadoop.hbase.conf.PropagatingConfigurationObserver;
-122import 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
-123import 
org.apache.hadoop.hbase.coprocessor.RegionObserver.MutationType;
-124import 
org.apache.hadoop.hbase.errorhandling.ForeignExceptionSnare;
-125import 
org.apache.hadoop.hbase.exceptions.FailedSanityCheckException;
-126import 
org.apache.hadoop.hbase.exceptions.TimeoutIOException;
-127import 
org.apache.hadoop.hbase.exceptions.UnknownProtocolException;
-128import 
org.apache.hadoop.hbase.filter.ByteArrayComparable;
-129import 
org.apache.hadoop.hbase.filter.FilterWrapper;
-130import 
org.apache.hadoop.hbase.filter.IncompatibleFilterException;
-131import 
org.apache.hadoop.hbase.io.HFileLink;
-132import 
org.apache.hadoop.hbase.io.HeapSize;
-133import 
org.apache.hadoop.hbase.io.TimeRange;
-134import 
org.apache.hadoop.hbase.io.hfile.BlockCache;
-135import 
org.apache.hadoop.hbase.io.hfile.HFile;
-136import 
org.apache.hadoop.hbase.ipc.CallerDisconnectedException;
-137import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
-138import 
org.apache.hadoop.hbase.ipc.RpcCall;
-139import 
org.apache.hadoop.hbase.ipc.RpcServer;
-140import 
org.apache.hadoop.hbase.mob.MobFileCache;
-141import 
org.apache.hadoop.hbase.monitoring.MonitoredTask;
-142import 
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-143import 
org.apache.hadoop.hbase.quotas.RegionServerSpaceQuotaManager;
-144import 
org.apache.hadoop.hbase.regionserver.MultiVersionConcurrencyControl.WriteEntry;
-145import 

[07/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.

2018-12-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html
index 79cb21b..d8d391b 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.EncodedScanner.html
@@ -378,1508 +378,1510 @@
 370
 371  @Override
 372  public void returnBlock(HFileBlock 
block) {
-373BlockCache blockCache = 
this.cacheConf.getBlockCache();
-374if (blockCache != null  
block != null) {
-375  BlockCacheKey cacheKey = new 
BlockCacheKey(this.getFileContext().getHFileName(),
-376  block.getOffset(), 
this.isPrimaryReplicaReader(), block.getBlockType());
-377  blockCache.returnBlock(cacheKey, 
block);
-378}
-379  }
-380  /**
-381   * @return the first key in the file. 
May be null if file has no entries. Note
-382   * that this is not the first 
row key, but rather the byte form of the
-383   * first KeyValue.
-384   */
-385  @Override
-386  public OptionalCell 
getFirstKey() {
-387if (dataBlockIndexReader == null) {
-388  throw new 
BlockIndexNotLoadedException();
-389}
-390return dataBlockIndexReader.isEmpty() 
? Optional.empty()
-391: 
Optional.of(dataBlockIndexReader.getRootBlockKey(0));
-392  }
-393
-394  /**
-395   * TODO left from {@link HFile} version 
1: move this to StoreFile after Ryan's
-396   * patch goes in to eliminate {@link 
KeyValue} here.
-397   *
-398   * @return the first row key, or null 
if the file is empty.
-399   */
-400  @Override
-401  public Optionalbyte[] 
getFirstRowKey() {
-402// We have to copy the row part to 
form the row key alone
-403return 
getFirstKey().map(CellUtil::cloneRow);
-404  }
-405
-406  /**
-407   * TODO left from {@link HFile} version 
1: move this to StoreFile after
-408   * Ryan's patch goes in to eliminate 
{@link KeyValue} here.
-409   *
-410   * @return the last row key, or null if 
the file is empty.
-411   */
-412  @Override
-413  public Optionalbyte[] 
getLastRowKey() {
-414// We have to copy the row part to 
form the row key alone
-415return 
getLastKey().map(CellUtil::cloneRow);
-416  }
-417
-418  /** @return number of KV entries in 
this HFile */
-419  @Override
-420  public long getEntries() {
-421return trailer.getEntryCount();
-422  }
-423
-424  /** @return comparator */
-425  @Override
-426  public CellComparator getComparator() 
{
-427return comparator;
-428  }
-429
-430  /** @return compression algorithm */
-431  @Override
-432  public Compression.Algorithm 
getCompressionAlgorithm() {
-433return compressAlgo;
-434  }
-435
-436  /**
-437   * @return the total heap size of data 
and meta block indexes in bytes. Does
-438   * not take into account 
non-root blocks of a multilevel data index.
-439   */
-440  @Override
-441  public long indexSize() {
-442return (dataBlockIndexReader != null 
? dataBlockIndexReader.heapSize() : 0)
-443+ ((metaBlockIndexReader != null) 
? metaBlockIndexReader.heapSize()
-444: 0);
-445  }
-446
-447  @Override
-448  public String getName() {
-449return name;
-450  }
-451
-452  @Override
-453  public HFileBlockIndex.BlockIndexReader 
getDataBlockIndexReader() {
-454return dataBlockIndexReader;
-455  }
-456
-457  @Override
-458  public FixedFileTrailer getTrailer() 
{
-459return trailer;
-460  }
-461
-462  @Override
-463  public boolean isPrimaryReplicaReader() 
{
-464return primaryReplicaReader;
-465  }
-466
-467  @Override
-468  public FileInfo loadFileInfo() throws 
IOException {
-469return fileInfo;
-470  }
-471
-472  /**
-473   * An exception thrown when an 
operation requiring a scanner to be seeked
-474   * is invoked on a scanner that is not 
seeked.
-475   */
-476  @SuppressWarnings("serial")
-477  public static class NotSeekedException 
extends IllegalStateException {
-478public NotSeekedException() {
-479  super("Not seeked to a 
key/value");
-480}
-481  }
-482
-483  protected static class HFileScannerImpl 
implements HFileScanner {
-484private ByteBuff blockBuffer;
-485protected final boolean 
cacheBlocks;
-486protected final boolean pread;
-487protected final boolean 
isCompaction;
-488private int currKeyLen;
-489private int currValueLen;
-490private int currMemstoreTSLen;
-491private long currMemstoreTS;
-492// Updated but never read?
-493protected AtomicInteger blockFetches 
= new AtomicInteger(0);
-494protected final HFile.Reader 
reader;
-495private int currTagsLen;
-496// buffer backed keyonlyKV
-497private ByteBufferKeyOnlyKeyValue 

[07/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.

2018-12-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
index a957d31..62f81b6 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
@@ -142,5192 +142,5186 @@
 134import org.apache.hadoop.hbase.wal.WAL;
 135import 
org.apache.hadoop.hbase.wal.WALFactory;
 136import 
org.apache.hadoop.hbase.wal.WALSplitter;
-137import 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator;
-138import 
org.apache.hadoop.hbase.zookeeper.ZKUtil;
-139import 
org.apache.hadoop.hbase.zookeeper.ZKWatcher;
-140import 
org.apache.hadoop.hbase.zookeeper.ZNodePaths;
-141import 
org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException;
-142import 
org.apache.hadoop.ipc.RemoteException;
-143import 
org.apache.hadoop.security.UserGroupInformation;
-144import 
org.apache.hadoop.util.ReflectionUtils;
-145import org.apache.hadoop.util.Tool;
-146import 
org.apache.hadoop.util.ToolRunner;
-147import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-148import 
org.apache.hbase.thirdparty.com.google.common.collect.Sets;
-149import 
org.apache.yetus.audience.InterfaceAudience;
-150import 
org.apache.yetus.audience.InterfaceStability;
-151import 
org.apache.zookeeper.KeeperException;
-152import org.slf4j.Logger;
-153import org.slf4j.LoggerFactory;
-154
-155import 
org.apache.hbase.thirdparty.com.google.common.base.Joiner;
-156import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
-157import 
org.apache.hbase.thirdparty.com.google.common.collect.ImmutableList;
-158import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
-159import 
org.apache.hbase.thirdparty.com.google.common.collect.Multimap;
-160import 
org.apache.hbase.thirdparty.com.google.common.collect.Ordering;
-161import 
org.apache.hbase.thirdparty.com.google.common.collect.TreeMultimap;
-162
-163import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-164import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface;
-165
-166/**
-167 * HBaseFsck (hbck) is a tool for 
checking and repairing region consistency and
-168 * table integrity problems in a 
corrupted HBase. This tool was written for hbase-1.x. It does not
-169 * work with hbase-2.x; it can read state 
but is not allowed to change state; i.e. effect 'repair'.
-170 * See hbck2 (HBASE-19121) for a hbck 
tool for hbase2.
-171 *
-172 * p
-173 * Region consistency checks verify that 
hbase:meta, region deployment on region
-174 * servers and the state of data in HDFS 
(.regioninfo files) all are in
-175 * accordance.
-176 * p
-177 * Table integrity checks verify that all 
possible row keys resolve to exactly
-178 * one region of a table.  This means 
there are no individual degenerate
-179 * or backwards regions; no holes between 
regions; and that there are no
-180 * overlapping regions.
-181 * p
-182 * The general repair strategy works in 
two phases:
-183 * ol
-184 * li Repair Table Integrity on 
HDFS. (merge or fabricate regions)
-185 * li Repair Region Consistency 
with hbase:meta and assignments
-186 * /ol
-187 * p
-188 * For table integrity repairs, the 
tables' region directories are scanned
-189 * for .regioninfo files.  Each table's 
integrity is then verified.  If there
-190 * are any orphan regions (regions with 
no .regioninfo files) or holes, new
-191 * regions are fabricated.  Backwards 
regions are sidelined as well as empty
-192 * degenerate (endkey==startkey) regions. 
 If there are any overlapping regions,
-193 * a new region is created and all data 
is merged into the new region.
-194 * p
-195 * Table integrity repairs deal solely 
with HDFS and could potentially be done
-196 * offline -- the hbase region servers or 
master do not need to be running.
-197 * This phase can eventually be used to 
completely reconstruct the hbase:meta table in
-198 * an offline fashion.
-199 * p
-200 * Region consistency requires three 
conditions -- 1) valid .regioninfo file
-201 * present in an HDFS region dir,  2) 
valid row with .regioninfo data in META,
-202 * and 3) a region is deployed only at 
the regionserver that was assigned to
-203 * with proper state in the master.
-204 * p
-205 * Region consistency repairs require 
hbase to be online so that hbck can
-206 * contact the HBase master and region 
servers.  The hbck#connect() method must
-207 * first be called successfully.  Much of 
the region consistency information
-208 * is transient and less risky to 
repair.
-209 * p
-210 * If hbck is run from 

[07/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.

2018-11-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html
index 8e3eb2f..3080df8 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.Builder.html
@@ -174,7 +174,7 @@
 166
 167private Builder(NamespaceDescriptor 
ns) {
 168  this.bName = ns.name;
-169  this.bConfiguration = 
ns.configuration;
+169  
this.bConfiguration.putAll(ns.configuration);
 170}
 171
 172private Builder(String name) {

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html
index 8e3eb2f..3080df8 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/NamespaceDescriptor.html
@@ -174,7 +174,7 @@
 166
 167private Builder(NamespaceDescriptor 
ns) {
 168  this.bName = ns.name;
-169  this.bConfiguration = 
ns.configuration;
+169  
this.bConfiguration.putAll(ns.configuration);
 170}
 171
 172private Builder(String name) {



[07/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.

2018-11-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/testdevapidocs/org/apache/hadoop/hbase/security/access/TestTablePermissions.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/security/access/TestTablePermissions.html
 
b/testdevapidocs/org/apache/hadoop/hbase/security/access/TestTablePermissions.html
index e9217bf..10e2501 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/security/access/TestTablePermissions.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/security/access/TestTablePermissions.html
@@ -224,11 +224,11 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 void
-checkMultimapEqual(org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.TablePermissionfirst,
-  
org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.TablePermissionsecond)
+checkMultimapEqual(org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.UserPermissionfirst,
+  
org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.UserPermissionsecond)
 
 
-private 
org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.TablePermission
+private 
org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.UserPermission
 createPermissions()
 
 
@@ -484,7 +484,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 testPersistence
-publicvoidtestPersistence()
+publicvoidtestPersistence()
  throws https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exception
 
 Throws:
@@ -498,7 +498,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 testSerialization
-publicvoidtestSerialization()
+publicvoidtestSerialization()
throws https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exception
 
 Throws:
@@ -512,7 +512,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 createPermissions
-privateorg.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.TablePermissioncreatePermissions()
+privateorg.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.UserPermissioncreatePermissions()
 
 
 
@@ -521,8 +521,8 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 checkMultimapEqual
-publicvoidcheckMultimapEqual(org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.TablePermissionfirst,
-   
org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.TablePermissionsecond)
+publicvoidcheckMultimapEqual(org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.UserPermissionfirst,
+   
org.apache.hbase.thirdparty.com.google.common.collect.ListMultimaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 

[07/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.

2018-10-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestRegionServerAbortTimeout.TestAbortTimeoutTask.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestRegionServerAbortTimeout.TestAbortTimeoutTask.html
 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestRegionServerAbortTimeout.TestAbortTimeoutTask.html
new file mode 100644
index 000..42d7bdd
--- /dev/null
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestRegionServerAbortTimeout.TestAbortTimeoutTask.html
@@ -0,0 +1,298 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+TestRegionServerAbortTimeout.TestAbortTimeoutTask (Apache HBase 
3.0.0-SNAPSHOT Test API)
+
+
+
+
+
+var methods = {"i0":10};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.regionserver
+Class TestRegionServerAbortTimeout.TestAbortTimeoutTask
+
+
+
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+
+
+https://docs.oracle.com/javase/8/docs/api/java/util/TimerTask.html?is-external=true;
 title="class or interface in java.util">java.util.TimerTask
+
+
+org.apache.hadoop.hbase.regionserver.TestRegionServerAbortTimeout.TestAbortTimeoutTask
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true;
 title="class or interface in java.lang">Runnable
+
+
+Enclosing class:
+TestRegionServerAbortTimeout
+
+
+
+static class TestRegionServerAbortTimeout.TestAbortTimeoutTask
+extends https://docs.oracle.com/javase/8/docs/api/java/util/TimerTask.html?is-external=true;
 title="class or interface in java.util">TimerTask
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+TestAbortTimeoutTask()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsInstance MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+void
+run()
+
+
+
+
+
+
+Methods inherited from classjava.util.https://docs.oracle.com/javase/8/docs/api/java/util/TimerTask.html?is-external=true;
 title="class or interface in java.util">TimerTask
+https://docs.oracle.com/javase/8/docs/api/java/util/TimerTask.html?is-external=true#cancel--;
 title="class or interface in java.util">cancel, https://docs.oracle.com/javase/8/docs/api/java/util/TimerTask.html?is-external=true#scheduledExecutionTime--;
 title="class or interface in java.util">scheduledExecutionTime
+
+
+
+
+
+Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
 title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
 title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
 title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/ja
 va/lang/Object.html?is-external=true#notifyAll--" title="class or interface in 
java.lang">notifyAll, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
 title="class or interface in java.lang">toString, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--;
 

[07/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.

2018-10-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.LeaseRecovery.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.LeaseRecovery.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.LeaseRecovery.html
index 9b964f6..98ef11a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.LeaseRecovery.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.LeaseRecovery.html
@@ -105,7 +105,7 @@
 097 * will first be initialized to the 
oldest file's tracker(which is stored in the trailer), using the
 098 * method {@link 
ProcedureStoreTracker#resetTo(ProcedureStoreTracker, boolean)}, and then merge 
it
 099 * with the tracker of every newer wal 
files, using the
-100 * {@link 
ProcedureStoreTracker#setDeletedIfModifiedInBoth(ProcedureStoreTracker, 
boolean)}.
+100 * {@link 
ProcedureStoreTracker#setDeletedIfModifiedInBoth(ProcedureStoreTracker)}.
 101 * If we find out
 102 * that all the modified procedures for 
the oldest wal file are modified or deleted in newer wal
 103 * files, then we can delete it. This is 
because that, every time we call
@@ -1181,244 +1181,243 @@
 1173}
 1174
 1175// compute the holding tracker.
-1176//  - the first WAL is used for the 
'updates'
-1177//  - the global tracker is passed 
in first to decide which procedures are not
-1178//exist anymore, so we can mark 
them as deleted in holdingCleanupTracker.
-1179//Only global tracker have the 
whole picture here.
-1180//  - the other WALs are scanned to 
remove procs already updated in a newer wal.
-1181//If it is updated in a newer 
wal, we can mark it as delelted in holdingCleanupTracker
-1182//But, we can not delete it if 
it was shown deleted in the newer wal, as said
-1183//above.
-1184// TODO: exit early if 
holdingCleanupTracker.isEmpty()
-1185
holdingCleanupTracker.resetTo(logs.getFirst().getTracker(), true);
-1186//Passing in the global tracker, we 
can delete the procedures not in the global
-1187//tracker, because they are deleted 
in the later logs
-1188
holdingCleanupTracker.setDeletedIfModifiedInBoth(storeTracker, true);
-1189for (int i = 1, size = logs.size() - 
1; i  size; ++i) {
-1190  // Set deleteIfNotExists to false 
since a single log's tracker is passed in.
-1191  // Since a specific procedure may 
not show up in the log at all(not executed or
-1192  // updated during the time), we 
can not delete the procedure just because this log
-1193  // don't have the info of the 
procedure. We can delete the procedure only if
-1194  // in this log's tracker, it was 
cleanly showed that the procedure is modified or deleted
-1195  // in the corresponding 
BitSetNode.
-1196  
holdingCleanupTracker.setDeletedIfModifiedInBoth(logs.get(i).getTracker(), 
false);
-1197}
-1198  }
-1199
-1200  /**
-1201   * Remove all logs with logId = 
{@code lastLogId}.
-1202   */
-1203  private void removeAllLogs(long 
lastLogId, String why) {
-1204if (logs.size() = 1) {
-1205  return;
-1206}
-1207
-1208LOG.info("Remove all state logs with 
ID less than {}, since {}", lastLogId, why);
-1209
-1210boolean removed = false;
-1211while (logs.size()  1) {
-1212  ProcedureWALFile log = 
logs.getFirst();
-1213  if (lastLogId  log.getLogId()) 
{
-1214break;
-1215  }
-1216  removeLogFile(log, 
walArchiveDir);
-1217  removed = true;
-1218}
-1219
-1220if (removed) {
-1221  buildHoldingCleanupTracker();
-1222}
-1223  }
-1224
-1225  private boolean removeLogFile(final 
ProcedureWALFile log, final Path walArchiveDir) {
-1226try {
-1227  LOG.trace("Removing log={}", 
log);
-1228  log.removeFile(walArchiveDir);
-1229  logs.remove(log);
-1230  LOG.debug("Removed log={}, 
activeLogs={}", log, logs);
-1231  assert logs.size()  0 : 
"expected at least one log";
-1232} catch (IOException e) {
-1233  LOG.error("Unable to remove log: " 
+ log, e);
-1234  return false;
-1235}
-1236return true;
-1237  }
-1238
-1239  // 
==
-1240  //  FileSystem Log Files helpers
-1241  // 
==
-1242  public Path getWALDir() {
-1243return this.walDir;
-1244  }
-1245
-1246  @VisibleForTesting
-1247  Path getWalArchiveDir() {
-1248return this.walArchiveDir;
-1249  }
-1250
-1251  public FileSystem getFileSystem() {
-1252return this.fs;
-1253  }
-1254
-1255  protected Path getLogFilePath(final 
long logId) throws IOException {
-1256return new Path(walDir, 
String.format(LOG_PREFIX + 

[07/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.

2018-10-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
index ed3db7a..156dabb 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.GetTillDoneOrException.html
@@ -5542,785 +5542,825 @@
 5534  }
 5535
 5536  @Test
-5537  public void testWriteRequestsCounter() 
throws IOException {
-5538byte[] fam = 
Bytes.toBytes("info");
-5539byte[][] families = { fam };
-5540this.region = initHRegion(tableName, 
method, CONF, families);
+5537  public void 
testReverseScanWhenPutCellsAfterOpenReverseScan() throws Exception {
+5538byte[] cf1 = Bytes.toBytes("CF1");
+5539byte[][] families = { cf1 };
+5540byte[] col = Bytes.toBytes("C");
 5541
-5542Assert.assertEquals(0L, 
region.getWriteRequestsCount());
-5543
-5544Put put = new Put(row);
-5545put.addColumn(fam, fam, fam);
-5546
-5547Assert.assertEquals(0L, 
region.getWriteRequestsCount());
-5548region.put(put);
-5549Assert.assertEquals(1L, 
region.getWriteRequestsCount());
-5550region.put(put);
-5551Assert.assertEquals(2L, 
region.getWriteRequestsCount());
-5552region.put(put);
-5553Assert.assertEquals(3L, 
region.getWriteRequestsCount());
-5554
-region.delete(new Delete(row));
-5556Assert.assertEquals(4L, 
region.getWriteRequestsCount());
-5557  }
-5558
-5559  @Test
-5560  public void 
testOpenRegionWrittenToWAL() throws Exception {
-5561final ServerName serverName = 
ServerName.valueOf(name.getMethodName(), 100, 42);
-5562final RegionServerServices rss = 
spy(TEST_UTIL.createMockRegionServerService(serverName));
-5563
-5564HTableDescriptor htd = new 
HTableDescriptor(TableName.valueOf(name.getMethodName()));
-5565htd.addFamily(new 
HColumnDescriptor(fam1));
-5566htd.addFamily(new 
HColumnDescriptor(fam2));
-5567
-5568HRegionInfo hri = new 
HRegionInfo(htd.getTableName(),
-5569  HConstants.EMPTY_BYTE_ARRAY, 
HConstants.EMPTY_BYTE_ARRAY);
-5570
-5571// open the region w/o rss and wal 
and flush some files
-5572region =
-5573 
HBaseTestingUtility.createRegionAndWAL(hri, TEST_UTIL.getDataTestDir(), 
TEST_UTIL
-5574 .getConfiguration(), 
htd);
-5575assertNotNull(region);
-5576
-5577// create a file in fam1 for the 
region before opening in OpenRegionHandler
-5578region.put(new 
Put(Bytes.toBytes("a")).addColumn(fam1, fam1, fam1));
-5579region.flush(true);
-5580
HBaseTestingUtility.closeRegionAndWAL(region);
+5542HBaseConfiguration conf = new 
HBaseConfiguration();
+5543this.region = initHRegion(tableName, 
method, conf, families);
+5544
+5545Put put = new 
Put(Bytes.toBytes("16"));
+5546put.addColumn(cf1, col, 
Bytes.toBytes("val"));
+5547region.put(put);
+5548Put put2 = new 
Put(Bytes.toBytes("15"));
+5549put2.addColumn(cf1, col, 
Bytes.toBytes("val"));
+5550region.put(put2);
+5551
+5552// Create a reverse scan
+5553Scan scan = new 
Scan(Bytes.toBytes("16"));
+5554scan.setReversed(true);
+RegionScannerImpl scanner = 
region.getScanner(scan);
+5556
+5557// Put a lot of cells that have 
sequenceIDs grater than the readPt of the reverse scan
+5558for (int i = 10; i  20; 
i++) {
+5559  Put p = new Put(Bytes.toBytes("" + 
i));
+5560  p.addColumn(cf1, col, 
Bytes.toBytes("" + i));
+5561  region.put(p);
+5562}
+5563ListCell currRow = new 
ArrayList();
+5564boolean hasNext;
+5565do {
+5566  hasNext = scanner.next(currRow);
+5567} while (hasNext);
+5568
+5569assertEquals(2, currRow.size());
+5570assertEquals("16", 
Bytes.toString(currRow.get(0).getRowArray(),
+5571  currRow.get(0).getRowOffset(), 
currRow.get(0).getRowLength()));
+5572assertEquals("15", 
Bytes.toString(currRow.get(1).getRowArray(),
+5573  currRow.get(1).getRowOffset(), 
currRow.get(1).getRowLength()));
+5574  }
+5575
+5576  @Test
+5577  public void testWriteRequestsCounter() 
throws IOException {
+5578byte[] fam = 
Bytes.toBytes("info");
+5579byte[][] families = { fam };
+5580this.region = initHRegion(tableName, 
method, CONF, families);
 5581
-5582ArgumentCaptorWALEdit 
editCaptor = ArgumentCaptor.forClass(WALEdit.class);
+5582Assert.assertEquals(0L, 
region.getWriteRequestsCount());
 5583
-5584// capture append() calls
-5585WAL wal = mockWAL();
-5586when(rss.getWAL((HRegionInfo) 
any())).thenReturn(wal);
-5587
-5588region = HRegion.openHRegion(hri, 
htd, 

[07/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.

2018-10-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/util/ByteBufferUtils.ConverterHolder.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/util/ByteBufferUtils.ConverterHolder.html 
b/devapidocs/org/apache/hadoop/hbase/util/ByteBufferUtils.ConverterHolder.html
new file mode 100644
index 000..5a1ec8a
--- /dev/null
+++ 
b/devapidocs/org/apache/hadoop/hbase/util/ByteBufferUtils.ConverterHolder.html
@@ -0,0 +1,348 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+ByteBufferUtils.ConverterHolder (Apache HBase 3.0.0-SNAPSHOT 
API)
+
+
+
+
+
+var methods = {"i0":9};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.util
+Class 
ByteBufferUtils.ConverterHolder
+
+
+
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.util.ByteBufferUtils.ConverterHolder
+
+
+
+
+
+
+
+Enclosing class:
+ByteBufferUtils
+
+
+
+static class ByteBufferUtils.ConverterHolder
+extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
+
+
+
+
+
+
+
+
+
+
+
+Nested Class Summary
+
+Nested Classes
+
+Modifier and Type
+Class and Description
+
+
+(package private) static class
+ByteBufferUtils.ConverterHolder.PureJavaConverter
+
+
+(package private) static class
+ByteBufferUtils.ConverterHolder.UnsafeConverter
+
+
+
+
+
+
+
+
+
+Field Summary
+
+Fields
+
+Modifier and Type
+Field and Description
+
+
+(package private) static ByteBufferUtils.Converter
+BEST_CONVERTER
+
+
+(package private) static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+UNSAFE_CONVERTER_NAME
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+ConverterHolder()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsStatic MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+(package private) static ByteBufferUtils.Converter
+getBestConverter()
+
+
+
+
+
+
+Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
 title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
 title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
 title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/ja
 va/lang/Object.html?is-external=true#notifyAll--" title="class or interface in 
java.lang">notifyAll, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
 title="class or interface in java.lang">toString, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--;
 title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-;
 title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-;
 title="class or interface in java.lang">wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Field Detail
+
+
+
+
+
+UNSAFE_CONVERTER_NAME
+static 

[07/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.

2018-10-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestSplitTableRegionProcedure.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestSplitTableRegionProcedure.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestSplitTableRegionProcedure.html
index 34e3682..da13fee 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestSplitTableRegionProcedure.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestSplitTableRegionProcedure.html
@@ -356,7 +356,7 @@
 348final TableName tableName = 
TableName.valueOf(name.getMethodName());
 349final 
ProcedureExecutorMasterProcedureEnv procExec = 
getMasterProcedureExecutor();
 350
-351RegionInfo [] regions = 
MasterProcedureTestingUtility.createTable(
+351RegionInfo[] regions = 
MasterProcedureTestingUtility.createTable(
 352  procExec, tableName, null, 
ColumnFamilyName1, ColumnFamilyName2);
 353insertData(tableName);
 354int splitRowNum = startRowNum + 
rowCount / 2;
@@ -374,181 +374,182 @@
 366long procId = 
procExec.submitProcedure(
 367  new 
SplitTableRegionProcedure(procExec.getEnvironment(), regions[0], splitKey));
 368
-369// Failing before 
SPLIT_TABLE_REGION_CREATE_DAUGHTER_REGIONS we should trigger the
+369// Failing before 
SPLIT_TABLE_REGION_UPDATE_META we should trigger the
 370// rollback
-371// NOTE: the 3 (number before 
SPLIT_TABLE_REGION_CREATE_DAUGHTER_REGIONS step) is
+371// NOTE: the 7 (number of 
SPLIT_TABLE_REGION_UPDATE_META step) is
 372// hardcoded, so you have to look at 
this test at least once when you add a new step.
-373int numberOfSteps = 3;
-374
MasterProcedureTestingUtility.testRollbackAndDoubleExecution(procExec, procId, 
numberOfSteps,
+373int lastStep = 7;
+374
MasterProcedureTestingUtility.testRollbackAndDoubleExecution(procExec, procId, 
lastStep,
 375true);
 376// check that we have only 1 region
 377assertEquals(1, 
UTIL.getAdmin().getRegions(tableName).size());
-378ListHRegion daughters = 
UTIL.getMiniHBaseCluster().getRegions(tableName);
-379assertEquals(1, daughters.size());
-380verifyData(daughters.get(0), 
startRowNum, rowCount,
-381Bytes.toBytes(ColumnFamilyName1), 
Bytes.toBytes(ColumnFamilyName2));
-382
-383assertEquals(splitSubmittedCount + 1, 
splitProcMetrics.getSubmittedCounter().getCount());
-384assertEquals(splitFailedCount + 1, 
splitProcMetrics.getFailedCounter().getCount());
-385  }
-386
-387  @Test
-388  public void 
testRecoveryAndDoubleExecution() throws Exception {
-389final TableName tableName = 
TableName.valueOf(name.getMethodName());
-390final 
ProcedureExecutorMasterProcedureEnv procExec = 
getMasterProcedureExecutor();
-391
-392RegionInfo [] regions = 
MasterProcedureTestingUtility.createTable(
-393  procExec, tableName, null, 
ColumnFamilyName1, ColumnFamilyName2);
-394insertData(tableName);
-395int splitRowNum = startRowNum + 
rowCount / 2;
-396byte[] splitKey = Bytes.toBytes("" + 
splitRowNum);
-397
-398assertTrue("not able to find a 
splittable region", regions != null);
-399assertTrue("not able to find a 
splittable region", regions.length == 1);
-400
ProcedureTestingUtility.waitNoProcedureRunning(procExec);
-401
ProcedureTestingUtility.setKillIfHasParent(procExec, false);
-402
ProcedureTestingUtility.setKillAndToggleBeforeStoreUpdate(procExec, true);
-403
-404// collect AM metrics before test
-405collectAssignmentManagerMetrics();
-406
-407// Split region of the table
-408long procId = 
procExec.submitProcedure(
-409  new 
SplitTableRegionProcedure(procExec.getEnvironment(), regions[0], splitKey));
-410
-411// Restart the executor and execute 
the step twice
-412
MasterProcedureTestingUtility.testRecoveryAndDoubleExecution(procExec, 
procId);
-413
ProcedureTestingUtility.assertProcNotFailed(procExec, procId);
-414
-415verify(tableName, splitRowNum);
-416
-417assertEquals(splitSubmittedCount + 1, 
splitProcMetrics.getSubmittedCounter().getCount());
-418assertEquals(splitFailedCount, 
splitProcMetrics.getFailedCounter().getCount());
-419  }
-420
-421  @Test
-422  public void testSplitWithoutPONR() 
throws Exception {
-423final TableName tableName = 
TableName.valueOf(name.getMethodName());
-424final 
ProcedureExecutorMasterProcedureEnv procExec = 
getMasterProcedureExecutor();
-425
-426RegionInfo [] regions = 
MasterProcedureTestingUtility.createTable(
-427procExec, tableName, null, 
ColumnFamilyName1, ColumnFamilyName2);
-428insertData(tableName);
-429int splitRowNum = startRowNum + 
rowCount / 2;
-430byte[] splitKey = Bytes.toBytes("" + 
splitRowNum);
-431
-432assertTrue("not able to find a 

[07/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.

2018-10-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.Strategies.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.Strategies.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.Strategies.html
index 2142742..be2a512 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.Strategies.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.Strategies.html
@@ -129,174 +129,171 @@
 121
 122  /** delegate provider for WAL 
creation/roll/close */
 123  public static final String 
DELEGATE_PROVIDER = "hbase.wal.regiongrouping.delegate.provider";
-124  public static final String 
DELEGATE_PROVIDER_CLASS =
-125
"hbase.wal.regiongrouping.delegate.provider.class";
-126  public static final String 
DEFAULT_DELEGATE_PROVIDER = WALFactory.Providers.defaultProvider
-127  .name();
+124  public static final String 
DEFAULT_DELEGATE_PROVIDER = WALFactory.Providers.defaultProvider
+125  .name();
+126
+127  private static final String 
META_WAL_GROUP_NAME = "meta";
 128
-129  private static final String 
META_WAL_GROUP_NAME = "meta";
-130
-131  /** A group-provider mapping, make sure 
one-one rather than many-one mapping */
-132  private final ConcurrentMapString, 
WALProvider cached = new ConcurrentHashMap();
+129  /** A group-provider mapping, make sure 
one-one rather than many-one mapping */
+130  private final ConcurrentMapString, 
WALProvider cached = new ConcurrentHashMap();
+131
+132  private final KeyLockerString 
createLock = new KeyLocker();
 133
-134  private final KeyLockerString 
createLock = new KeyLocker();
-135
-136  private RegionGroupingStrategy 
strategy;
-137  private WALFactory factory;
-138  private Configuration conf;
-139  private ListWALActionsListener 
listeners = new ArrayList();
-140  private String providerId;
-141  private Class? extends 
WALProvider providerClass;
-142
-143  @Override
-144  public void init(WALFactory factory, 
Configuration conf, String providerId) throws IOException {
-145if (null != strategy) {
-146  throw new 
IllegalStateException("WALProvider.init should only be called once.");
-147}
-148this.conf = conf;
-149this.factory = factory;
-150StringBuilder sb = new 
StringBuilder().append(factory.factoryId);
-151if (providerId != null) {
-152  if 
(providerId.startsWith(WAL_FILE_NAME_DELIMITER)) {
-153sb.append(providerId);
-154  } else {
-155
sb.append(WAL_FILE_NAME_DELIMITER).append(providerId);
-156  }
-157}
-158this.providerId = sb.toString();
-159this.strategy = getStrategy(conf, 
REGION_GROUPING_STRATEGY, DEFAULT_REGION_GROUPING_STRATEGY);
-160this.providerClass = 
factory.getProviderClass(DELEGATE_PROVIDER_CLASS, DELEGATE_PROVIDER,
-161DEFAULT_DELEGATE_PROVIDER);
-162  }
-163
-164  private WALProvider 
createProvider(String group) throws IOException {
-165WALProvider provider = 
WALFactory.createProvider(providerClass);
-166provider.init(factory, conf,
-167  
META_WAL_PROVIDER_ID.equals(providerId) ? META_WAL_PROVIDER_ID : group);
-168provider.addWALActionsListener(new 
MetricsWAL());
-169return provider;
-170  }
-171
-172  @Override
-173  public ListWAL getWALs() {
-174return 
cached.values().stream().flatMap(p - 
p.getWALs().stream()).collect(Collectors.toList());
-175  }
-176
-177  private WAL getWAL(String group) throws 
IOException {
-178WALProvider provider = 
cached.get(group);
-179if (provider == null) {
-180  Lock lock = 
createLock.acquireLock(group);
-181  try {
-182provider = cached.get(group);
-183if (provider == null) {
-184  provider = 
createProvider(group);
-185  
listeners.forEach(provider::addWALActionsListener);
-186  cached.put(group, provider);
-187}
-188  } finally {
-189lock.unlock();
-190  }
-191}
-192return provider.getWAL(null);
-193  }
-194
-195  @Override
-196  public WAL getWAL(RegionInfo region) 
throws IOException {
-197String group;
-198if 
(META_WAL_PROVIDER_ID.equals(this.providerId)) {
-199  group = META_WAL_GROUP_NAME;
-200} else {
-201  byte[] id;
-202  byte[] namespace;
-203  if (region != null) {
-204id = 
region.getEncodedNameAsBytes();
-205namespace = 
region.getTable().getNamespace();
-206  } else {
-207id = 
HConstants.EMPTY_BYTE_ARRAY;
-208namespace = null;
-209  }
-210  group = strategy.group(id, 
namespace);
-211}
-212return getWAL(group);
-213  }
-214
-215  @Override
-216  public void shutdown() throws 
IOException {
-217// save the last exception and 
rethrow
-218IOException failure = null;
-219for (WALProvider provider: 
cached.values()) {
-220  

[07/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.

2018-09-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/testdevapidocs/org/apache/hadoop/hbase/package-use.html
--
diff --git a/testdevapidocs/org/apache/hadoop/hbase/package-use.html 
b/testdevapidocs/org/apache/hadoop/hbase/package-use.html
index 2ae4566..29060fa 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/package-use.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/package-use.html
@@ -279,194 +279,198 @@
 
 
 
-org.apache.hadoop.hbase.master.snapshot
+org.apache.hadoop.hbase.master.replication
 
 
 
-org.apache.hadoop.hbase.metrics
+org.apache.hadoop.hbase.master.snapshot
 
 
 
-org.apache.hadoop.hbase.metrics.impl
+org.apache.hadoop.hbase.metrics
 
 
 
-org.apache.hadoop.hbase.mob
+org.apache.hadoop.hbase.metrics.impl
 
 
 
-org.apache.hadoop.hbase.mob.compactions
+org.apache.hadoop.hbase.mob
 
 
 
-org.apache.hadoop.hbase.monitoring
+org.apache.hadoop.hbase.mob.compactions
 
 
 
-org.apache.hadoop.hbase.mttr
+org.apache.hadoop.hbase.monitoring
 
 
 
-org.apache.hadoop.hbase.namespace
+org.apache.hadoop.hbase.mttr
 
 
 
-org.apache.hadoop.hbase.net
+org.apache.hadoop.hbase.namespace
 
 
 
-org.apache.hadoop.hbase.nio
+org.apache.hadoop.hbase.net
 
 
 
-org.apache.hadoop.hbase.procedure
+org.apache.hadoop.hbase.nio
 
 
 
-org.apache.hadoop.hbase.procedure2
+org.apache.hadoop.hbase.procedure
 
 
 
-org.apache.hadoop.hbase.procedure2.store
+org.apache.hadoop.hbase.procedure2
 
 
 
-org.apache.hadoop.hbase.procedure2.store.wal
+org.apache.hadoop.hbase.procedure2.store
 
 
 
-org.apache.hadoop.hbase.procedure2.util
+org.apache.hadoop.hbase.procedure2.store.wal
 
 
 
-org.apache.hadoop.hbase.quotas
+org.apache.hadoop.hbase.procedure2.util
 
 
 
-org.apache.hadoop.hbase.quotas.policies
+org.apache.hadoop.hbase.quotas
 
 
 
-org.apache.hadoop.hbase.regionserver
+org.apache.hadoop.hbase.quotas.policies
 
 
 
-org.apache.hadoop.hbase.regionserver.compactions
+org.apache.hadoop.hbase.regionserver
 
 
 
-org.apache.hadoop.hbase.regionserver.querymatcher
+org.apache.hadoop.hbase.regionserver.compactions
 
 
 
-org.apache.hadoop.hbase.regionserver.throttle
+org.apache.hadoop.hbase.regionserver.querymatcher
 
 
 
-org.apache.hadoop.hbase.regionserver.wal
+org.apache.hadoop.hbase.regionserver.throttle
 
 
 
-org.apache.hadoop.hbase.replication
+org.apache.hadoop.hbase.regionserver.wal
 
 
 
-org.apache.hadoop.hbase.replication.master
+org.apache.hadoop.hbase.replication
 
 
 
-org.apache.hadoop.hbase.replication.multiwal
+org.apache.hadoop.hbase.replication.master
 
 
 
-org.apache.hadoop.hbase.replication.regionserver
+org.apache.hadoop.hbase.replication.multiwal
 
 
 
-org.apache.hadoop.hbase.rest
+org.apache.hadoop.hbase.replication.regionserver
 
 
 
-org.apache.hadoop.hbase.rest.client
+org.apache.hadoop.hbase.rest
 
 
 
-org.apache.hadoop.hbase.rest.model
+org.apache.hadoop.hbase.rest.client
 
 
 
-org.apache.hadoop.hbase.rsgroup
+org.apache.hadoop.hbase.rest.model
 
 
 
-org.apache.hadoop.hbase.security
+org.apache.hadoop.hbase.rsgroup
 
 
 
-org.apache.hadoop.hbase.security.access
+org.apache.hadoop.hbase.security
 
 
 
-org.apache.hadoop.hbase.security.token
+org.apache.hadoop.hbase.security.access
 
 
 
-org.apache.hadoop.hbase.security.visibility
+org.apache.hadoop.hbase.security.token
 
 
 
-org.apache.hadoop.hbase.snapshot
+org.apache.hadoop.hbase.security.visibility
 
 
 
-org.apache.hadoop.hbase.test
+org.apache.hadoop.hbase.snapshot
 
 
 
-org.apache.hadoop.hbase.thrift
+org.apache.hadoop.hbase.test
 
 
 
-org.apache.hadoop.hbase.thrift2
+org.apache.hadoop.hbase.thrift
 
 
 
-org.apache.hadoop.hbase.tool
+org.apache.hadoop.hbase.thrift2
 
 
 
-org.apache.hadoop.hbase.tool.coprocessor
+org.apache.hadoop.hbase.tool
 
 
 
-org.apache.hadoop.hbase.trace
+org.apache.hadoop.hbase.tool.coprocessor
 
 
 
-org.apache.hadoop.hbase.types
+org.apache.hadoop.hbase.trace
 
 
 
-org.apache.hadoop.hbase.util
+org.apache.hadoop.hbase.types
 
 
 
-org.apache.hadoop.hbase.util.compaction
+org.apache.hadoop.hbase.util
 
 
 
-org.apache.hadoop.hbase.util.hbck
+org.apache.hadoop.hbase.util.compaction
 
 
 
-org.apache.hadoop.hbase.wal
+org.apache.hadoop.hbase.util.hbck
 
 
 
-org.apache.hadoop.hbase.zookeeper
+org.apache.hadoop.hbase.wal
 
 
 
+org.apache.hadoop.hbase.zookeeper
+
+
+
 org.apache.hbase.archetypes.exemplars.client
 
 
@@ -1802,6 +1806,28 @@
 
 
 
+
+
+
+
+Classes in org.apache.hadoop.hbase
 used by org.apache.hadoop.hbase.master.replication
+
+Class and Description
+
+
+
+HBaseClassTestRule
+The class level TestRule for all the tests.
+
+
+
+HBaseTestingUtility
+Facility for testing HBase.
+
+
+
+
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/testdevapidocs/org/apache/hadoop/hbase/procedure/TestProcedureDescriber.TestProcedure.html
--
diff --git 

[07/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.

2018-09-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
index 566f410..da040ad 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
@@ -341,8361 +341,8425 @@
 333  private final int 
rowLockWaitDuration;
 334  static final int 
DEFAULT_ROWLOCK_WAIT_DURATION = 3;
 335
-336  // The internal wait duration to 
acquire a lock before read/update
-337  // from the region. It is not per row. 
The purpose of this wait time
-338  // is to avoid waiting a long time 
while the region is busy, so that
-339  // we can release the IPC handler soon 
enough to improve the
-340  // availability of the region server. 
It can be adjusted by
-341  // tuning configuration 
"hbase.busy.wait.duration".
-342  final long busyWaitDuration;
-343  static final long 
DEFAULT_BUSY_WAIT_DURATION = HConstants.DEFAULT_HBASE_RPC_TIMEOUT;
-344
-345  // If updating multiple rows in one 
call, wait longer,
-346  // i.e. waiting for busyWaitDuration * 
# of rows. However,
-347  // we can limit the max multiplier.
-348  final int maxBusyWaitMultiplier;
-349
-350  // Max busy wait duration. There is no 
point to wait longer than the RPC
-351  // purge timeout, when a RPC call will 
be terminated by the RPC engine.
-352  final long maxBusyWaitDuration;
-353
-354  // Max cell size. If nonzero, the 
maximum allowed size for any given cell
-355  // in bytes
-356  final long maxCellSize;
-357
-358  // Number of mutations for minibatch 
processing.
-359  private final int miniBatchSize;
+336  private Path regionDir;
+337  private FileSystem walFS;
+338
+339  // The internal wait duration to 
acquire a lock before read/update
+340  // from the region. It is not per row. 
The purpose of this wait time
+341  // is to avoid waiting a long time 
while the region is busy, so that
+342  // we can release the IPC handler soon 
enough to improve the
+343  // availability of the region server. 
It can be adjusted by
+344  // tuning configuration 
"hbase.busy.wait.duration".
+345  final long busyWaitDuration;
+346  static final long 
DEFAULT_BUSY_WAIT_DURATION = HConstants.DEFAULT_HBASE_RPC_TIMEOUT;
+347
+348  // If updating multiple rows in one 
call, wait longer,
+349  // i.e. waiting for busyWaitDuration * 
# of rows. However,
+350  // we can limit the max multiplier.
+351  final int maxBusyWaitMultiplier;
+352
+353  // Max busy wait duration. There is no 
point to wait longer than the RPC
+354  // purge timeout, when a RPC call will 
be terminated by the RPC engine.
+355  final long maxBusyWaitDuration;
+356
+357  // Max cell size. If nonzero, the 
maximum allowed size for any given cell
+358  // in bytes
+359  final long maxCellSize;
 360
-361  // negative number indicates infinite 
timeout
-362  static final long 
DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L;
-363  final ExecutorService 
rowProcessorExecutor = Executors.newCachedThreadPool();
-364
-365  private final 
ConcurrentHashMapRegionScanner, Long scannerReadPoints;
-366
-367  /**
-368   * The sequence ID that was 
enLongAddered when this region was opened.
-369   */
-370  private long openSeqNum = 
HConstants.NO_SEQNUM;
-371
-372  /**
-373   * The default setting for whether to 
enable on-demand CF loading for
-374   * scan requests to this region. 
Requests can override it.
-375   */
-376  private boolean 
isLoadingCfsOnDemandDefault = false;
-377
-378  private final AtomicInteger 
majorInProgress = new AtomicInteger(0);
-379  private final AtomicInteger 
minorInProgress = new AtomicInteger(0);
+361  // Number of mutations for minibatch 
processing.
+362  private final int miniBatchSize;
+363
+364  // negative number indicates infinite 
timeout
+365  static final long 
DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L;
+366  final ExecutorService 
rowProcessorExecutor = Executors.newCachedThreadPool();
+367
+368  private final 
ConcurrentHashMapRegionScanner, Long scannerReadPoints;
+369
+370  /**
+371   * The sequence ID that was 
enLongAddered when this region was opened.
+372   */
+373  private long openSeqNum = 
HConstants.NO_SEQNUM;
+374
+375  /**
+376   * The default setting for whether to 
enable on-demand CF loading for
+377   * scan requests to this region. 
Requests can override it.
+378   */
+379  private boolean 
isLoadingCfsOnDemandDefault = false;
 380
-381  //
-382  // Context: During replay we want to 
ensure that we do not lose any data. So, we
-383  // have to be conservative in how we 
replay wals. For each store, we calculate
-384  // the maxSeqId up to which the store 
was flushed. And, skip the 

[07/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.

2018-09-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html
index acc491f..e6c6561 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html
@@ -26,256 +26,255 @@
 018 */
 019package 
org.apache.hadoop.hbase.regionserver;
 020
-021import java.io.IOException;
-022import java.util.Collection;
-023import java.util.List;
-024import java.util.Map.Entry;
-025import 
java.util.concurrent.ConcurrentMap;
-026
+021import com.google.protobuf.Service;
+022import java.io.IOException;
+023import java.util.Collection;
+024import java.util.List;
+025import java.util.Map.Entry;
+026import 
java.util.concurrent.ConcurrentMap;
 027import 
org.apache.hadoop.hbase.Abortable;
 028import org.apache.hadoop.hbase.Server;
-029import 
org.apache.hadoop.hbase.TableName;
-030import 
org.apache.hadoop.hbase.client.RegionInfo;
-031import 
org.apache.hadoop.hbase.client.locking.EntityLock;
-032import 
org.apache.hadoop.hbase.executor.ExecutorService;
-033import 
org.apache.hadoop.hbase.ipc.RpcServerInterface;
-034import 
org.apache.hadoop.hbase.quotas.RegionServerRpcQuotaManager;
-035import 
org.apache.hadoop.hbase.quotas.RegionServerSpaceQuotaManager;
-036import 
org.apache.hadoop.hbase.quotas.RegionSizeStore;
-037import 
org.apache.hadoop.hbase.regionserver.compactions.CompactionRequester;
-038import 
org.apache.hadoop.hbase.regionserver.throttle.ThroughputController;
-039import org.apache.hadoop.hbase.wal.WAL;
-040import 
org.apache.yetus.audience.InterfaceAudience;
-041import 
org.apache.zookeeper.KeeperException;
+029import 
org.apache.hadoop.hbase.TableDescriptors;
+030import 
org.apache.hadoop.hbase.TableName;
+031import 
org.apache.hadoop.hbase.client.RegionInfo;
+032import 
org.apache.hadoop.hbase.client.locking.EntityLock;
+033import 
org.apache.hadoop.hbase.executor.ExecutorService;
+034import 
org.apache.hadoop.hbase.ipc.RpcServerInterface;
+035import 
org.apache.hadoop.hbase.quotas.RegionServerRpcQuotaManager;
+036import 
org.apache.hadoop.hbase.quotas.RegionServerSpaceQuotaManager;
+037import 
org.apache.hadoop.hbase.quotas.RegionSizeStore;
+038import 
org.apache.hadoop.hbase.regionserver.compactions.CompactionRequester;
+039import 
org.apache.hadoop.hbase.regionserver.throttle.ThroughputController;
+040import org.apache.hadoop.hbase.wal.WAL;
+041import 
org.apache.yetus.audience.InterfaceAudience;
 042
 043import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode;
 044
-045import com.google.protobuf.Service;
-046
-047/**
-048 * A curated subset of services provided 
by {@link HRegionServer}.
-049 * For use internally only. Passed to 
Managers, Services and Chores so can pass less-than-a
-050 * full-on HRegionServer at test-time. Be 
judicious adding API. Changes cause ripples through
-051 * the code base.
-052 */
-053@InterfaceAudience.Private
-054public interface RegionServerServices 
extends Server, MutableOnlineRegions, FavoredNodesForRegion {
-055
-056  /** @return the WAL for a particular 
region. Pass null for getting the
-057   * default (common) WAL */
-058  WAL getWAL(RegionInfo regionInfo) 
throws IOException;
-059
-060  /** @return the List of WALs that are 
used by this server
-061   *  Doesn't include the meta WAL
-062   */
-063  ListWAL getWALs() throws 
IOException;
-064
-065  /**
-066   * @return Implementation of {@link 
FlushRequester} or null. Usually it will not be null unless
-067   * during intialization.
-068   */
-069  FlushRequester getFlushRequester();
-070
-071  /**
-072   * @return Implementation of {@link 
CompactionRequester} or null. Usually it will not be null
-073   * unless during 
intialization.
-074   */
-075  CompactionRequester 
getCompactionRequestor();
-076
-077  /**
-078   * @return the RegionServerAccounting 
for this Region Server
-079   */
-080  RegionServerAccounting 
getRegionServerAccounting();
-081
-082  /**
-083   * @return RegionServer's instance of 
{@link RegionServerRpcQuotaManager}
-084   */
-085  RegionServerRpcQuotaManager 
getRegionServerRpcQuotaManager();
-086
-087  /**
-088   * @return RegionServer's instance of 
{@link SecureBulkLoadManager}
-089   */
-090  SecureBulkLoadManager 
getSecureBulkLoadManager();
-091
-092  /**
-093   * @return RegionServer's instance of 
{@link RegionServerSpaceQuotaManager}
-094   */
-095  RegionServerSpaceQuotaManager 
getRegionServerSpaceQuotaManager();
-096
-097  /**
-098   * Context 

[07/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.

2018-09-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
index 2c14c50..43c66a8 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
@@ -46,2104 +46,2113 @@
 038import 
java.util.concurrent.atomic.AtomicLong;
 039import java.util.stream.Collectors;
 040import java.util.stream.Stream;
-041import 
org.apache.hadoop.conf.Configuration;
-042import 
org.apache.hadoop.hbase.HConstants;
-043import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
-044import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-045import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
-046import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
-047import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
-048import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
-049import 
org.apache.hadoop.hbase.security.User;
-050import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-051import 
org.apache.hadoop.hbase.util.IdLock;
-052import 
org.apache.hadoop.hbase.util.NonceKey;
-053import 
org.apache.hadoop.hbase.util.Threads;
-054import 
org.apache.yetus.audience.InterfaceAudience;
-055import org.slf4j.Logger;
-056import org.slf4j.LoggerFactory;
-057
-058import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-059import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
-060
-061import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
-062
-063/**
-064 * Thread Pool that executes the 
submitted procedures.
-065 * The executor has a ProcedureStore 
associated.
-066 * Each operation is logged and on 
restart the pending procedures are resumed.
-067 *
-068 * Unless the Procedure code throws an 
error (e.g. invalid user input)
-069 * the procedure will complete (at some 
point in time), On restart the pending
-070 * procedures are resumed and the once 
failed will be rolledback.
-071 *
-072 * The user can add procedures to the 
executor via submitProcedure(proc)
-073 * check for the finished state via 
isFinished(procId)
-074 * and get the result via 
getResult(procId)
-075 */
-076@InterfaceAudience.Private
-077public class 
ProcedureExecutorTEnvironment {
-078  private static final Logger LOG = 
LoggerFactory.getLogger(ProcedureExecutor.class);
-079
-080  public static final String 
CHECK_OWNER_SET_CONF_KEY = "hbase.procedure.check.owner.set";
-081  private static final boolean 
DEFAULT_CHECK_OWNER_SET = false;
-082
-083  public static final String 
WORKER_KEEP_ALIVE_TIME_CONF_KEY =
-084  
"hbase.procedure.worker.keep.alive.time.msec";
-085  private static final long 
DEFAULT_WORKER_KEEP_ALIVE_TIME = TimeUnit.MINUTES.toMillis(1);
-086
-087  /**
-088   * {@link #testing} is non-null when 
ProcedureExecutor is being tested. Tests will try to
-089   * break PE having it fail at various 
junctures. When non-null, testing is set to an instance of
-090   * the below internal {@link Testing} 
class with flags set for the particular test.
-091   */
-092  Testing testing = null;
-093
-094  /**
-095   * Class with parameters describing how 
to fail/die when in testing-context.
-096   */
-097  public static class Testing {
-098protected boolean killIfHasParent = 
true;
-099protected boolean killIfSuspended = 
false;
-100
-101/**
-102 * Kill the PE BEFORE we store state 
to the WAL. Good for figuring out if a Procedure is
-103 * persisting all the state it needs 
to recover after a crash.
-104 */
-105protected boolean 
killBeforeStoreUpdate = false;
-106protected boolean 
toggleKillBeforeStoreUpdate = false;
-107
-108/**
-109 * Set when we want to fail AFTER 
state has been stored into the WAL. Rarely used. HBASE-20978
-110 * is about a case where memory-state 
was being set after store to WAL where a crash could
-111 * cause us to get stuck. This flag 
allows killing at what was a vulnerable time.
-112 */
-113protected boolean 
killAfterStoreUpdate = false;
-114protected boolean 
toggleKillAfterStoreUpdate = false;
-115
-116protected boolean 
shouldKillBeforeStoreUpdate() {
-117  final boolean kill = 
this.killBeforeStoreUpdate;
-118  if 
(this.toggleKillBeforeStoreUpdate) {
-119this.killBeforeStoreUpdate = 
!kill;
-120LOG.warn("Toggle KILL before 
store update to: " + this.killBeforeStoreUpdate);
-121  }
-122  return kill;
-123}
-124
-125protected boolean 
shouldKillBeforeStoreUpdate(boolean isSuspended, boolean hasParent) {

[07/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.

2018-09-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.Cluster.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.Cluster.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.Cluster.html
index c372545..af3b364 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.Cluster.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.Cluster.html
@@ -1279,322 +1279,339 @@
 1271ListRegionInfo 
lastFewRegions = new ArrayList();
 1272// assign the remaining by going 
through the list and try to assign to servers one-by-one
 1273int serverIdx = 
RANDOM.nextInt(numServers);
-1274for (RegionInfo region : 
unassignedRegions) {
+1274OUTER : for (RegionInfo region : 
unassignedRegions) {
 1275  boolean assigned = false;
-1276  for (int j = 0; j  numServers; 
j++) { // try all servers one by one
+1276  INNER : for (int j = 0; j  
numServers; j++) { // try all servers one by one
 1277ServerName serverName = 
servers.get((j + serverIdx) % numServers);
 1278if 
(!cluster.wouldLowerAvailability(region, serverName)) {
 1279  ListRegionInfo 
serverRegions =
 1280  
assignments.computeIfAbsent(serverName, k - new ArrayList());
-1281  serverRegions.add(region);
-1282  cluster.doAssignRegion(region, 
serverName);
-1283  serverIdx = (j + serverIdx + 
1) % numServers; //remain from next server
-1284  assigned = true;
-1285  break;
-1286}
-1287  }
-1288  if (!assigned) {
-1289lastFewRegions.add(region);
-1290  }
-1291}
-1292// just sprinkle the rest of the 
regions on random regionservers. The balanceCluster will
-1293// make it optimal later. we can end 
up with this if numReplicas  numServers.
-1294for (RegionInfo region : 
lastFewRegions) {
-1295  int i = 
RANDOM.nextInt(numServers);
-1296  ServerName server = 
servers.get(i);
-1297  ListRegionInfo 
serverRegions = assignments.computeIfAbsent(server, k - new 
ArrayList());
-1298  serverRegions.add(region);
-1299  cluster.doAssignRegion(region, 
server);
-1300}
-1301return assignments;
-1302  }
-1303
-1304  protected Cluster 
createCluster(ListServerName servers, CollectionRegionInfo 
regions) {
-1305// Get the snapshot of the current 
assignments for the regions in question, and then create
-1306// a cluster out of it. Note that we 
might have replicas already assigned to some servers
-1307// earlier. So we want to get the 
snapshot to see those assignments, but this will only contain
-1308// replicas of the regions that are 
passed (for performance).
-1309MapServerName, 
ListRegionInfo clusterState = 
getRegionAssignmentsByServer(regions);
-1310
-1311for (ServerName server : servers) 
{
-1312  if 
(!clusterState.containsKey(server)) {
-1313clusterState.put(server, 
EMPTY_REGION_LIST);
-1314  }
-1315}
-1316return new Cluster(regions, 
clusterState, null, this.regionFinder,
-1317rackManager);
-1318  }
-1319
-1320  private ListServerName 
findIdleServers(ListServerName servers) {
-1321return 
this.services.getServerManager()
-1322
.getOnlineServersListWithPredicator(servers, IDLE_SERVER_PREDICATOR);
-1323  }
-1324
-1325  /**
-1326   * Used to assign a single region to a 
random server.
-1327   */
-1328  @Override
-1329  public ServerName 
randomAssignment(RegionInfo regionInfo, ListServerName servers)
-1330  throws HBaseIOException {
-1331
metricsBalancer.incrMiscInvocations();
-1332if (servers != null  
servers.contains(masterServerName)) {
-1333  if (shouldBeOnMaster(regionInfo)) 
{
-1334return masterServerName;
-1335  }
-1336  if 
(!LoadBalancer.isTablesOnMaster(getConf())) {
-1337// Guarantee we do not put any 
regions on master
-1338servers = new 
ArrayList(servers);
-1339
servers.remove(masterServerName);
-1340  }
-1341}
-1342
-1343int numServers = servers == null ? 0 
: servers.size();
-1344if (numServers == 0) {
-1345  LOG.warn("Wanted to retain 
assignment but no servers to assign to");
-1346  return null;
-1347}
-1348if (numServers == 1) { // Only one 
server, nothing fancy we can do here
-1349  return servers.get(0);
-1350}
-1351ListServerName idleServers = 
findIdleServers(servers);
-1352if (idleServers.size() == 1) {
-1353  return idleServers.get(0);
-1354}
-1355final ListServerName 
finalServers = idleServers.isEmpty() ?
-1356servers : idleServers;
-1357ListRegionInfo regions = 
Lists.newArrayList(regionInfo);
-1358Cluster cluster = 
createCluster(finalServers, regions);
-1359return 

[07/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.

2018-08-31 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
index d11176a..2c14c50 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
@@ -982,1050 +982,1168 @@
 974  }
 975
 976  /**
-977   * Add a new root-procedure to the 
executor.
-978   * @param proc the new procedure to 
execute.
-979   * @param nonceKey the registered 
unique identifier for this operation from the client or process.
-980   * @return the procedure id, that can 
be used to monitor the operation
-981   */
-982  
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="NP_NULL_ON_SOME_PATH",
-983  justification = "FindBugs is blind 
to the check-for-null")
-984  public long 
submitProcedure(ProcedureTEnvironment proc, NonceKey nonceKey) {
-985
Preconditions.checkArgument(lastProcId.get() = 0);
-986
-987prepareProcedure(proc);
-988
-989final Long currentProcId;
-990if (nonceKey != null) {
-991  currentProcId = 
nonceKeysToProcIdsMap.get(nonceKey);
-992  
Preconditions.checkArgument(currentProcId != null,
-993"Expected nonceKey=" + nonceKey + 
" to be reserved, use registerNonce(); proc=" + proc);
-994} else {
-995  currentProcId = nextProcId();
-996}
-997
-998// Initialize the procedure
-999proc.setNonceKey(nonceKey);
-1000
proc.setProcId(currentProcId.longValue());
-1001
-1002// Commit the transaction
-1003store.insert(proc, null);
-1004LOG.debug("Stored {}", proc);
-1005
-1006// Add the procedure to the 
executor
-1007return pushProcedure(proc);
-1008  }
-1009
-1010  /**
-1011   * Add a set of new root-procedure to 
the executor.
-1012   * @param procs the new procedures to 
execute.
-1013   */
-1014  // TODO: Do we need to take nonces 
here?
-1015  public void 
submitProcedures(ProcedureTEnvironment[] procs) {
-1016
Preconditions.checkArgument(lastProcId.get() = 0);
-1017if (procs == null || procs.length 
= 0) {
-1018  return;
-1019}
-1020
-1021// Prepare procedure
-1022for (int i = 0; i  procs.length; 
++i) {
-1023  
prepareProcedure(procs[i]).setProcId(nextProcId());
-1024}
-1025
-1026// Commit the transaction
-1027store.insert(procs);
-1028if (LOG.isDebugEnabled()) {
-1029  LOG.debug("Stored " + 
Arrays.toString(procs));
-1030}
-1031
-1032// Add the procedure to the 
executor
-1033for (int i = 0; i  procs.length; 
++i) {
-1034  pushProcedure(procs[i]);
-1035}
-1036  }
-1037
-1038  private ProcedureTEnvironment 
prepareProcedure(ProcedureTEnvironment proc) {
-1039
Preconditions.checkArgument(proc.getState() == ProcedureState.INITIALIZING);
-1040
Preconditions.checkArgument(!proc.hasParent(), "unexpected parent", proc);
-1041if (this.checkOwnerSet) {
-1042  
Preconditions.checkArgument(proc.hasOwner(), "missing owner");
-1043}
-1044return proc;
-1045  }
-1046
-1047  private long 
pushProcedure(ProcedureTEnvironment proc) {
-1048final long currentProcId = 
proc.getProcId();
+977   * Bypass a procedure. If the procedure 
is set to bypass, all the logic in
+978   * execute/rollback will be ignored and 
it will return success, whatever.
+979   * It is used to recover buggy stuck 
procedures, releasing the lock resources
+980   * and letting other procedures to run. 
Bypassing one procedure (and its ancestors will
+981   * be bypassed automatically) may leave 
the cluster in a middle state, e.g. region
+982   * not assigned, or some hdfs files 
left behind. After getting rid of those stuck procedures,
+983   * the operators may have to do some 
clean up on hdfs or schedule some assign procedures
+984   * to let region online. DO AT YOUR OWN 
RISK.
+985   * p
+986   * A procedure can be bypassed only 
if
+987   * 1. The procedure is in state of 
RUNNABLE, WAITING, WAITING_TIMEOUT
+988   * or it is a root procedure without 
any child.
+989   * 2. No other worker thread is 
executing it
+990   * 3. No child procedure has been 
submitted
+991   *
+992   * p
+993   * If all the requirements are meet, 
the procedure and its ancestors will be
+994   * bypassed and persisted to WAL.
+995   *
+996   * p
+997   * If the procedure is in WAITING 
state, will set it to RUNNABLE add it to run queue.
+998   * TODO: What about WAITING_TIMEOUT?
+999   * @param id the procedure id
+1000   * @param lockWait time to wait lock
+1001   * @param force if force set to true, 
we will bypass the procedure even if it is executing.
+1002   *  This is for procedures 

[07/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.

2018-08-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/testdevapidocs/org/apache/hadoop/hbase/class-use/StartMiniClusterOption.Builder.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/class-use/StartMiniClusterOption.Builder.html
 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/StartMiniClusterOption.Builder.html
new file mode 100644
index 000..b378060
--- /dev/null
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/StartMiniClusterOption.Builder.html
@@ -0,0 +1,205 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+Uses of Class org.apache.hadoop.hbase.StartMiniClusterOption.Builder 
(Apache HBase 3.0.0-SNAPSHOT Test API)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+
+
+
+Uses of 
Classorg.apache.hadoop.hbase.StartMiniClusterOption.Builder
+
+
+
+
+
+Packages that use StartMiniClusterOption.Builder
+
+Package
+Description
+
+
+
+org.apache.hadoop.hbase
+
+
+
+
+
+
+
+
+
+
+Uses of StartMiniClusterOption.Builder in 
org.apache.hadoop.hbase
+
+Methods in org.apache.hadoop.hbase
 that return StartMiniClusterOption.Builder
+
+Modifier and Type
+Method and Description
+
+
+
+static StartMiniClusterOption.Builder
+StartMiniClusterOption.builder()
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.createRootDir(booleancreateRootDir)
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.createWALDir(booleancreateWALDir)
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.dataNodeHosts(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String[]dataNodeHosts)
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.masterClass(https://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? extends 
org.apache.hadoop.hbase.master.HMastermasterClass)
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.numDataNodes(intnumDataNodes)
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.numMasters(intnumMasters)
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.numRegionServers(intnumRegionServers)
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.numZkServers(intnumZkServers)
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.rsClass(https://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? extends MiniHBaseCluster.MiniHBaseClusterRegionServerrsClass)
+
+
+StartMiniClusterOption.Builder
+StartMiniClusterOption.Builder.rsPorts(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in 
java.lang">IntegerrsPorts)
+
+
+
+
+
+
+
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+
+
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+
+

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/testdevapidocs/org/apache/hadoop/hbase/class-use/StartMiniClusterOption.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/class-use/StartMiniClusterOption.html 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/StartMiniClusterOption.html
new file mode 100644
index 000..6972e77
--- /dev/null
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/StartMiniClusterOption.html
@@ -0,0 +1,186 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+Uses of Class org.apache.hadoop.hbase.StartMiniClusterOption (Apache 
HBase 3.0.0-SNAPSHOT Test API)
+
+
+
+
+
-var methods = {"i0":10,"i1":10,"i2":9,"i3":10,"i4":10};
+var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":9,"i5":10,"i6":10};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -120,7 +120,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-class AccessControlFilter
+class AccessControlFilter
 extends FilterBase
 NOTE: for internal use only by AccessController 
implementation
 
@@ -268,29 +268,37 @@ extends Method and Description
 
 
+boolean
+equals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objectobj)
+
+
 Filter.ReturnCode
 filterCell(Cellcell)
 A way to filter based on the column family, column 
qualifier and/or the column value.
 
 
-
+
 boolean
 filterRowKey(Cellcell)
 Filters a row based on the row key.
 
 
-
+
+int
+hashCode()
+
+
 static AccessControlFilter
 parseFrom(byte[]pbBytes)
 
-
+
 void
 reset()
 Filters that are purely stateless and do nothing in their 
reset() methods can inherit
  this null/empty implementation.
 
 
-
+
 byte[]
 toByteArray()
 Return length 0 byte array for Filters that don't require 
special serialization
@@ -316,7 +324,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
-https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
 title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
 title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
 title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/ja
 va/lang/Object.html?is-external=true#notifyAll--" title="class or interface in 
java.lang">notifyAll, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--;
 title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-;
 title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-;
 title="class or interface in java.lang">wait
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
 title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
 title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notifyAll--;
 title="class or interface in java.lang">notifyAll, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--;
 title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.htm
 l?is-external=true#wait-long-" title="class or interface in 
java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-;
 title="class or interface in java.lang">wait
 
 
 
@@ -338,7 +346,7 @@ extends 
 
 authManager
-privateTableAuthManager 
authManager
+privateTableAuthManager 
authManager
 
 
 
@@ -347,7 +355,7 @@ extends 
 
 table
-privateTableName table
+privateTableName table
 
 
 
@@ -356,7 +364,7 @@ extends 
 
 user
-privateUser user
+privateUser user
 
 
 
@@ -365,7 +373,7 @@ extends 
 
 

[07/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
index 522d316..2dd2c14 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseTestingUtility.html
@@ -2380,54 +2380,58 @@
 TestServerCrashProcedure.util
 
 
+private static HBaseTestingUtility
+TestCreateTableProcedureMuitipleRegions.UTIL
+
+
 protected static HBaseTestingUtility
 TestTableDDLProcedureBase.UTIL
 
-
+
 protected static HBaseTestingUtility
 MasterProcedureSchedulerPerformanceEvaluation.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestCreateNamespaceProcedure.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestModifyNamespaceProcedure.UTIL
 
-
+
 private static HBaseTestingUtility
 TestProcedurePriority.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestMasterProcedureEvents.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestDeleteNamespaceProcedure.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestMasterObserverPostCalls.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestWALProcedureStoreOnHDFS.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestSafemodeBringsDownMaster.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestMasterFailoverWithProcedures.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestMasterProcedureWalLease.UTIL
 
-
+
 protected static HBaseTestingUtility
 TestProcedureAdmin.UTIL
 
@@ -2836,274 +2840,278 @@
 TestNewVersionBehaviorFromClientSide.TEST_UTIL
 
 
+protected static HBaseTestingUtility
+TestRecoveredEditsReplayAndAbort.TEST_UTIL
+
+
 private static HBaseTestingUtility
 TestCompactSplitThread.TEST_UTIL
 
-
+
 (package private) static HBaseTestingUtility
 TestRegionMergeTransactionOnCluster.TEST_UTIL
 
-
+
 protected static HBaseTestingUtility
 TestCompactionPolicy.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestWalAndCompactingMemStoreFlush.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestWALMonotonicallyIncreasingSeqId.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestDeleteMobTable.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestBlocksScanned.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestCompoundBloomFilter.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestMultiColumnScanner.TEST_UTIL
 
-
+
 protected static HBaseTestingUtility
 TestHRegion.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestScannerWithBulkload.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestHMobStore.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestStoreFileScannerWithTagCompression.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestScannerWithCorruptHFile.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestEndToEndSplitTransaction.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestScannerHeartbeatMessages.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestIsDeleteFailure.TEST_UTIL
 
-
+
 private HBaseTestingUtility
 TestStoreFileRefresherChore.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestRegionMove.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestRegionServerMetrics.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestCacheOnWriteInSchema.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestSettingTimeoutOnBlockingPoint.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestHRegionReplayEvents.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestRemoveRegionMetrics.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestRecoveredEdits.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestHRegionOnCluster.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestHStore.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestSplitLogWorker.TEST_UTIL
 
-
+
 private HBaseTestingUtility
 TestRegionServerHostname.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestEncryptionRandomKeying.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestPerColumnFamilyFlush.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestSeekOptimizations.TEST_UTIL
 
-
+
 (package private) HBaseTestingUtility
 TestReversibleScanners.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestMasterAddressTracker.TEST_UTIL
 
-
+
 private HBaseTestingUtility
 TestAtomicOperation.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestHdfsSnapshotHRegion.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestRegionServerReadRequestMetrics.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestTimestampFilterSeekHint.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 TestScanWithBloomError.TEST_UTIL
 
-
+
 private HBaseTestingUtility
 TestClusterId.TEST_UTIL
 
-
+
 private static HBaseTestingUtility
 

[07/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.

2018-08-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.RemoteProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.RemoteProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.RemoteProcedure.html
index 5c6ff8e..dbe617a 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.RemoteProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.RemoteProcedure.html
@@ -101,7 +101,7 @@ var activeTableTab = "activeTableTab";
 
 
 All Known Implementing Classes:
-AssignProcedure, RefreshPeerProcedure, RegionTransitionProcedure, SyncReplicationReplayWALRemoteProcedure,
 UnassignProcedure
+AssignProcedure, CloseRegionProcedure, OpenRegionProcedure, RefreshPeerProcedure, RegionRemoteProcedureBase, RegionTransitionProcedure, SyncReplicationReplayWALRemoteProcedure,
 UnassignProcedure
 
 
 Enclosing class:

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.LockState.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.LockState.html
 
b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.LockState.html
index 37edbe6..cad236d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.LockState.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.LockState.html
@@ -122,10 +122,6 @@
 protected Procedure.LockState
 MergeTableRegionsProcedure.acquireLock(MasterProcedureEnvenv)
 
-
-protected Procedure.LockState
-RegionTransitionProcedure.acquireLock(MasterProcedureEnvenv)
-
 
 
 
@@ -160,35 +156,29 @@
 
 
 protected Procedure.LockState
-RecoverMetaProcedure.acquireLock(MasterProcedureEnvenv)
-Deprecated.
-
-
-
-protected Procedure.LockState
 CreateTableProcedure.acquireLock(MasterProcedureEnvenv)
 
-
+
 protected Procedure.LockState
 AbstractStateMachineRegionProcedure.acquireLock(MasterProcedureEnvenv)
 
-
+
 protected Procedure.LockState
 CreateNamespaceProcedure.acquireLock(MasterProcedureEnvenv)
 
-
+
 protected Procedure.LockState
 AbstractStateMachineTableProcedure.acquireLock(MasterProcedureEnvenv)
 
-
+
 protected Procedure.LockState
 ServerCrashProcedure.acquireLock(MasterProcedureEnvenv)
 
-
+
 protected Procedure.LockState
 AbstractStateMachineNamespaceProcedure.acquireLock(MasterProcedureEnvenv)
 
-
+
 protected Procedure.LockState
 InitMetaProcedure.acquireLock(MasterProcedureEnvenv)
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html 
b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
index 776026f..937e9c2 100644
--- a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
+++ b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
@@ -164,6 +164,20 @@
 
 
 
+private void
+HMaster.checkUnsupportedProcedure(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? extends Procedure,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListProcedureMasterProcedureEnvprocsByType)
+In HBASE-20811, we have introduced a new TRSP to 
assign/unassign/move regions, and it is
+ incompatible with the old 
AssignProcedure/UnassignProcedure/MoveRegionProcedure.
+
+
+
+private void
+HMaster.checkUnsupportedProcedure(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? extends Procedure,https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListProcedureMasterProcedureEnvprocsByType)
+In HBASE-20811, we have introduced a new TRSP to 
assign/unassign/move regions, and it is
+ incompatible with the old 
AssignProcedure/UnassignProcedure/MoveRegionProcedure.
+
+
+
 void
 MasterCoprocessorHost.postGetProcedures(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListProcedure?procInfoList)
 
@@ -188,49 +202,81 @@
 
 class
 AssignProcedure
-Procedure that describe the assignment of a single 
region.
+Deprecated.
+Do not 

[07/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.

2018-08-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.WriterAndPath.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.WriterAndPath.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.WriterAndPath.html
index 63e4b46..514f830 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.WriterAndPath.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.WriterAndPath.html
@@ -468,15 +468,15 @@
 460   * creating it if necessary.
 461   * @param logEntry
 462   * @param fileNameBeingSplit the file 
being split currently. Used to generate tmp file name.
-463   * @param conf
-464   * @return Path to file into which to 
dump split log edits.
-465   * @throws IOException
-466   */
-467  @SuppressWarnings("deprecation")
-468  @VisibleForTesting
-469  static Path 
getRegionSplitEditsPath(final Entry logEntry, String fileNameBeingSplit,
-470  Configuration conf)
-471  throws IOException {
+463   * @param tmpDirName of the directory 
used to sideline old recovered edits file
+464   * @param conf
+465   * @return Path to file into which to 
dump split log edits.
+466   * @throws IOException
+467   */
+468  @SuppressWarnings("deprecation")
+469  @VisibleForTesting
+470  static Path 
getRegionSplitEditsPath(final Entry logEntry, String fileNameBeingSplit,
+471  String tmpDirName, Configuration 
conf) throws IOException {
 472FileSystem fs = 
FileSystem.get(conf);
 473Path rootDir = 
FSUtils.getRootDir(conf);
 474Path tableDir = 
FSUtils.getTableDir(rootDir, logEntry.getKey().getTableName());
@@ -491,7 +491,7 @@
 483  return null;
 484}
 485if (fs.exists(dir)  
fs.isFile(dir)) {
-486  Path tmp = new Path("/tmp");
+486  Path tmp = new Path(tmpDirName);
 487  if (!fs.exists(tmp)) {
 488fs.mkdirs(tmp);
 489  }
@@ -1520,411 +1520,413 @@
 1512 * @return a path with a write for 
that path. caller should close.
 1513 */
 1514WriterAndPath createWAP(byte[] 
region, Entry entry) throws IOException {
-1515  Path regionedits = 
getRegionSplitEditsPath(entry,
-1516  
fileBeingSplit.getPath().getName(), conf);
-1517  if (regionedits == null) {
-1518return null;
-1519  }
-1520  FileSystem rootFs = 
FileSystem.get(conf);
-1521  if (rootFs.exists(regionedits)) 
{
-1522LOG.warn("Found old edits file. 
It could be the "
-1523+ "result of a previous 
failed split attempt. Deleting " + regionedits + ", length="
-1524+ 
rootFs.getFileStatus(regionedits).getLen());
-1525if (!rootFs.delete(regionedits, 
false)) {
-1526  LOG.warn("Failed delete of old 
{}", regionedits);
-1527}
-1528  }
-1529  Writer w = 
createWriter(regionedits);
-1530  LOG.debug("Creating writer 
path={}", regionedits);
-1531  return new 
WriterAndPath(regionedits, w, entry.getKey().getSequenceId());
-1532}
-1533
-1534void filterCellByStore(Entry 
logEntry) {
-1535  Mapbyte[], Long 
maxSeqIdInStores =
-1536  
regionMaxSeqIdInStores.get(Bytes.toString(logEntry.getKey().getEncodedRegionName()));
-1537  if 
(MapUtils.isEmpty(maxSeqIdInStores)) {
-1538return;
-1539  }
-1540  // Create the array list for the 
cells that aren't filtered.
-1541  // We make the assumption that 
most cells will be kept.
-1542  ArrayListCell keptCells = 
new ArrayList(logEntry.getEdit().getCells().size());
-1543  for (Cell cell : 
logEntry.getEdit().getCells()) {
-1544if 
(CellUtil.matchingFamily(cell, WALEdit.METAFAMILY)) {
-1545  keptCells.add(cell);
-1546} else {
-1547  byte[] family = 
CellUtil.cloneFamily(cell);
-1548  Long maxSeqId = 
maxSeqIdInStores.get(family);
-1549  // Do not skip cell even if 
maxSeqId is null. Maybe we are in a rolling upgrade,
-1550  // or the master was crashed 
before and we can not get the information.
-1551  if (maxSeqId == null || 
maxSeqId.longValue()  logEntry.getKey().getSequenceId()) {
-1552keptCells.add(cell);
-1553  }
-1554}
-1555  }
-1556
-1557  // Anything in the keptCells array 
list is still live.
-1558  // So rather than removing the 
cells from the array list
-1559  // which would be an O(n^2) 
operation, we just replace the list
-1560  
logEntry.getEdit().setCells(keptCells);
-1561}
-1562
-1563@Override
-1564public void append(RegionEntryBuffer 
buffer) throws IOException {
-1565  appendBuffer(buffer, true);
-1566}
-1567
-1568WriterAndPath 
appendBuffer(RegionEntryBuffer buffer, boolean reusable) throws IOException{
-1569  ListEntry entries = 
buffer.entryBuffer;
-1570  if (entries.isEmpty()) {
-1571LOG.warn("got an empty buffer, 
skipping");
-1572return 

[07/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.

2018-08-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/plugins.html
--
diff --git a/plugins.html b/plugins.html
index 133655d..d889203 100644
--- a/plugins.html
+++ b/plugins.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Project Plugins
 
@@ -375,7 +375,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-01
+  Last Published: 
2018-08-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/poweredbyhbase.html
--
diff --git a/poweredbyhbase.html b/poweredbyhbase.html
index a8bae52..7a002ac 100644
--- a/poweredbyhbase.html
+++ b/poweredbyhbase.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Powered By Apache HBase™
 
@@ -769,7 +769,7 @@ under the License. -->
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-01
+  Last Published: 
2018-08-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/project-info.html
--
diff --git a/project-info.html b/project-info.html
index 95c4bbe..c3c3dc8 100644
--- a/project-info.html
+++ b/project-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Project Information
 
@@ -335,7 +335,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-01
+  Last Published: 
2018-08-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/project-reports.html
--
diff --git a/project-reports.html b/project-reports.html
index ed5b94c..c891241 100644
--- a/project-reports.html
+++ b/project-reports.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Generated Reports
 
@@ -305,7 +305,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-01
+  Last Published: 
2018-08-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/project-summary.html
--
diff --git a/project-summary.html b/project-summary.html
index d2d95c0..fd407ec 100644
--- a/project-summary.html
+++ b/project-summary.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Project Summary
 
@@ -331,7 +331,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-01
+  Last Published: 
2018-08-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/pseudo-distributed.html
--
diff --git a/pseudo-distributed.html b/pseudo-distributed.html
index 8263a90..66f093c 100644
--- a/pseudo-distributed.html
+++ b/pseudo-distributed.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase   
 Running Apache HBase (TM) in pseudo-distributed mode
@@ -308,7 +308,7 @@ under the License. -->
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-01
+  Last Published: 
2018-08-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/replication.html
--
diff --git a/replication.html b/replication.html
index cca75ea..c69b6039 100644
--- a/replication.html
+++ b/replication.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  
   Apache HBase (TM) Replication
@@ -303,7 +303,7 @@ under the License. -->
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-08-01
+  Last Published: 
2018-08-02
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/resources.html
--

[07/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.

2018-07-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
index 95f2a65..073d0d0 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
@@ -931,7 +931,7 @@
 923InitMetaProcedure initMetaProc = 
null;
 924if 
(assignmentManager.getRegionStates().getRegionState(RegionInfoBuilder.FIRST_META_REGIONINFO)
 925  .isOffline()) {
-926  OptionalProcedure? 
optProc = procedureExecutor.getProcedures().stream()
+926  
OptionalProcedureMasterProcedureEnv optProc = 
procedureExecutor.getProcedures().stream()
 927.filter(p - p instanceof 
InitMetaProcedure).findAny();
 928  if (optProc.isPresent()) {
 929initMetaProc = 
(InitMetaProcedure) optProc.get();
@@ -3210,566 +3210,567 @@
 3202  cpHost.preGetProcedures();
 3203}
 3204
-3205final ListProcedure? 
procList = this.procedureExecutor.getProcedures();
-3206
-3207if (cpHost != null) {
-3208  
cpHost.postGetProcedures(procList);
-3209}
-3210
-3211return procList;
-3212  }
-3213
-3214  @Override
-3215  public ListLockedResource 
getLocks() throws IOException {
-3216if (cpHost != null) {
-3217  cpHost.preGetLocks();
-3218}
-3219
-3220MasterProcedureScheduler 
procedureScheduler =
-3221  
procedureExecutor.getEnvironment().getProcedureScheduler();
-3222
-3223final ListLockedResource 
lockedResources = procedureScheduler.getLocks();
-3224
-3225if (cpHost != null) {
-3226  
cpHost.postGetLocks(lockedResources);
-3227}
-3228
-3229return lockedResources;
-3230  }
-3231
-3232  /**
-3233   * Returns the list of table 
descriptors that match the specified request
-3234   * @param namespace the namespace to 
query, or null if querying for all
-3235   * @param regex The regular expression 
to match against, or null if querying for all
-3236   * @param tableNameList the list of 
table names, or null if querying for all
-3237   * @param includeSysTables False to 
match only against userspace tables
-3238   * @return the list of table 
descriptors
-3239   */
-3240  public ListTableDescriptor 
listTableDescriptors(final String namespace, final String regex,
-3241  final ListTableName 
tableNameList, final boolean includeSysTables)
-3242  throws IOException {
-3243ListTableDescriptor htds = 
new ArrayList();
-3244if (cpHost != null) {
-3245  
cpHost.preGetTableDescriptors(tableNameList, htds, regex);
-3246}
-3247htds = getTableDescriptors(htds, 
namespace, regex, tableNameList, includeSysTables);
-3248if (cpHost != null) {
-3249  
cpHost.postGetTableDescriptors(tableNameList, htds, regex);
-3250}
-3251return htds;
-3252  }
-3253
-3254  /**
-3255   * Returns the list of table names 
that match the specified request
-3256   * @param regex The regular expression 
to match against, or null if querying for all
-3257   * @param namespace the namespace to 
query, or null if querying for all
-3258   * @param includeSysTables False to 
match only against userspace tables
-3259   * @return the list of table names
-3260   */
-3261  public ListTableName 
listTableNames(final String namespace, final String regex,
-3262  final boolean includeSysTables) 
throws IOException {
-3263ListTableDescriptor htds = 
new ArrayList();
-3264if (cpHost != null) {
-3265  cpHost.preGetTableNames(htds, 
regex);
-3266}
-3267htds = getTableDescriptors(htds, 
namespace, regex, null, includeSysTables);
-3268if (cpHost != null) {
-3269  cpHost.postGetTableNames(htds, 
regex);
-3270}
-3271ListTableName result = new 
ArrayList(htds.size());
-3272for (TableDescriptor htd: htds) 
result.add(htd.getTableName());
-3273return result;
-3274  }
-3275
-3276  /**
-3277   * @return list of table table 
descriptors after filtering by regex and whether to include system
-3278   *tables, etc.
-3279   * @throws IOException
-3280   */
-3281  private ListTableDescriptor 
getTableDescriptors(final ListTableDescriptor htds,
-3282  final String namespace, final 
String regex, final ListTableName tableNameList,
-3283  final boolean includeSysTables)
-3284  throws IOException {
-3285if (tableNameList == null || 
tableNameList.isEmpty()) {
-3286  // request for all 
TableDescriptors
-3287  CollectionTableDescriptor 
allHtds;
-3288  if (namespace != null  
namespace.length()  0) {
-3289// Do a check on the namespace 
existence. Will fail if does not exist.
-3290
this.clusterSchemaService.getNamespace(namespace);
-3291

[07/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.

2018-07-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.WriteRequestCostFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.WriteRequestCostFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.WriteRequestCostFunction.html
index 233dba3..91b9055 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.WriteRequestCostFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.WriteRequestCostFunction.html
@@ -540,1205 +540,1204 @@
 532  
sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) - {
 533DequeBalancerRegionLoad 
rLoads = oldLoads.get(Bytes.toString(regionName));
 534if (rLoads == null) {
-535  // There was nothing there
-536  rLoads = new 
ArrayDeque();
-537} else if (rLoads.size() = 
numRegionLoadsToRemember) {
-538  rLoads.remove();
-539}
-540rLoads.add(new 
BalancerRegionLoad(rm));
-541
loads.put(Bytes.toString(regionName), rLoads);
-542  });
-543});
-544
-545for(CostFromRegionLoadFunction cost : 
regionLoadFunctions) {
-546  cost.setLoads(loads);
-547}
-548  }
-549
-550  protected void initCosts(Cluster 
cluster) {
-551for (CostFunction c:costFunctions) 
{
-552  c.init(cluster);
-553}
-554  }
-555
-556  protected void 
updateCostsWithAction(Cluster cluster, Action action) {
-557for (CostFunction c : costFunctions) 
{
-558  c.postAction(action);
-559}
-560  }
-561
-562  /**
-563   * Get the names of the cost 
functions
-564   */
-565  public String[] getCostFunctionNames() 
{
-566if (costFunctions == null) return 
null;
-567String[] ret = new 
String[costFunctions.length];
-568for (int i = 0; i  
costFunctions.length; i++) {
-569  CostFunction c = 
costFunctions[i];
-570  ret[i] = 
c.getClass().getSimpleName();
-571}
-572
-573return ret;
-574  }
-575
-576  /**
-577   * This is the main cost function.  It 
will compute a cost associated with a proposed cluster
-578   * state.  All different costs will be 
combined with their multipliers to produce a double cost.
-579   *
-580   * @param cluster The state of the 
cluster
-581   * @param previousCost the previous 
cost. This is used as an early out.
-582   * @return a double of a cost 
associated with the proposed cluster state.  This cost is an
-583   * aggregate of all individual 
cost functions.
-584   */
-585  protected double computeCost(Cluster 
cluster, double previousCost) {
-586double total = 0;
-587
-588for (int i = 0; i  
costFunctions.length; i++) {
-589  CostFunction c = 
costFunctions[i];
-590  this.tempFunctionCosts[i] = 0.0;
-591
-592  if (c.getMultiplier() = 0) {
-593continue;
-594  }
-595
-596  Float multiplier = 
c.getMultiplier();
-597  Double cost = c.cost();
-598
-599  this.tempFunctionCosts[i] = 
multiplier*cost;
-600  total += 
this.tempFunctionCosts[i];
-601
-602  if (total  previousCost) {
-603break;
-604  }
-605}
-606
-607return total;
-608  }
-609
-610  /** Generates a candidate action to be 
applied to the cluster for cost function search */
-611  abstract static class 
CandidateGenerator {
-612abstract Cluster.Action 
generate(Cluster cluster);
-613
-614/**
-615 * From a list of regions pick a 
random one. Null can be returned which
-616 * {@link 
StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region 
move
-617 * rather than swap.
-618 *
-619 * @param clusterThe state of 
the cluster
-620 * @param server index of the 
server
-621 * @param chanceOfNoSwap Chance that 
this will decide to try a move rather
-622 *   than a 
swap.
-623 * @return a random {@link 
RegionInfo} or null if an asymmetrical move is
-624 * suggested.
-625 */
-626protected int 
pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) {
-627  // Check to see if this is just a 
move.
-628  if 
(cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat()  
chanceOfNoSwap) {
-629// signal a move only.
-630return -1;
-631  }
-632  int rand = 
RANDOM.nextInt(cluster.regionsPerServer[server].length);
-633  return 
cluster.regionsPerServer[server][rand];
-634
-635}
-636protected int 
pickRandomServer(Cluster cluster) {
-637  if (cluster.numServers  1) {
-638return -1;
-639  }
-640
-641  return 
RANDOM.nextInt(cluster.numServers);
-642}
-643
-644protected int pickRandomRack(Cluster 
cluster) {
-645  if (cluster.numRacks  1) {
-646return -1;
-647  

[07/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.

2018-07-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html 
b/apidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html
index 315944d..0c69e4c 100644
--- a/apidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html
+++ b/apidocs/org/apache/hadoop/hbase/class-use/CompareOperator.html
@@ -1,10 +1,10 @@
 http://www.w3.org/TR/html4/loose.dtd;>
 
-
+
 
 
 
-类 org.apache.hadoop.hbase.CompareOperator的使用 (Apache HBase 
3.0.0-SNAPSHOT API)
+Uses of Class org.apache.hadoop.hbase.CompareOperator (Apache HBase 
3.0.0-SNAPSHOT API)
 
 
 
@@ -12,7 +12,7 @@
 
 
 
-您的浏览器已禁用 JavaScript。
+JavaScript is disabled on your browser.
 
 
 
 
 
-跳过导航链接
+Skip navigation links
 
 
 
-
-概览
-程序包
-ç±»
-使用
-树
-已过时
-索引
-帮助
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
 
 
 
 
-上一个
-下一个
+Prev
+Next
 
 
-框架
-无框架
+Frames
+NoFrames
 
 
-所有类
+AllClasses
 
 
 

[07/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.

2018-07-19 Thread zhangduo
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html 
b/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
index 6b2e30c..dae577c 100644
--- a/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
+++ b/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
@@ -1,10 +1,10 @@
 http://www.w3.org/TR/html4/loose.dtd;>
 
-
+
 
 
 
-Uses of Interface org.apache.hadoop.hbase.CellScanner (Apache HBase 
3.0.0-SNAPSHOT API)
+接口 org.apache.hadoop.hbase.CellScanner的使用 (Apache HBase 
3.0.0-SNAPSHOT API)
 
 
 
@@ -12,7 +12,7 @@
 
 
 
-JavaScript is disabled on your browser.
+您的浏览器已禁用 JavaScript。
 
 
 
 
 
-Skip navigation links
+跳过导航链接
 
 
 
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
+
+概览
+程序包
+ç±»
+使用
+树
+已过时
+索引
+帮助
 
 
 
 
-Prev
-Next
+上一个
+下一个
 
 
-Frames
-NoFrames
+框架
+无框架
 
 
-AllClasses
+所有类
 
 
 
 
-

Uses of Interface
org.apache.hadoop.hbase.CellScanner

+

接口的使用
org.apache.hadoop.hbase.CellScanner

  • - - +
    Packages that use CellScanner 
    + - - + + @@ -89,7 +89,17 @@ @@ -100,37 +110,37 @@
  • -

    Uses of CellScanner in org.apache.hadoop.hbase

    -
  • 使用CellScanner的程序包  
    PackageDescription程序包说明
    org.apache.hadoop.hbase.client -
    Provides HBase Client
    +
    Provides HBase Client + +Table of Contents + + Overview +Example API Usage + + + Overview + To administer HBase, create and drop tables, list and alter tables, + use Admin.
    - +

    org.apache.hadoop.hbase中CellScanner的使用

    +
    Methods in org.apache.hadoop.hbase that return CellScanner 
    +
    返回

    [07/51] [partial] hbase-site git commit: Published site at 0f23784182ab88649de340d75804e0ff20dcd0fc.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
    index f118c08..cd9722f 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
    @@ -206,140 +206,175 @@
     198   */
     199  public static int 
    validateBlockAddition(Cacheable existing, Cacheable newBlock,
     200  
    BlockCacheKey cacheKey) {
    -201int comparison = 
    compareCacheBlock(existing, newBlock, true);
    +201int comparison = 
    compareCacheBlock(existing, newBlock, false);
     202if (comparison != 0) {
    -203  LOG.warn("Cached block contents 
    differ, trying to just compare the block contents " +
    -204  "without the next block. 
    CacheKey: " + cacheKey);
    -205
    -206  // compare the contents, if they 
    are not equal, we are in big trouble
    -207  int 
    comparisonWithoutNextBlockMetadata = compareCacheBlock(existing, newBlock, 
    false);
    -208
    -209  if 
    (comparisonWithoutNextBlockMetadata != 0) {
    -210throw new 
    RuntimeException("Cached block contents differ, which should not have 
    happened."
    -211+ "cacheKey:" + cacheKey);
    -212  }
    -213}
    -214return comparison;
    -215  }
    -216
    -217  /**
    -218   * Use one of these to keep a running 
    account of cached blocks by file.  Throw it away when done.
    -219   * This is different than metrics in 
    that it is stats on current state of a cache.
    -220   * See getLoadedCachedBlocksByFile
    -221   */
    -222  
    @JsonIgnoreProperties({"cachedBlockStatsByFile"})
    -223  public static class CachedBlocksByFile 
    {
    -224private int count;
    -225private int dataBlockCount;
    -226private long size;
    -227private long dataSize;
    -228private final long now = 
    System.nanoTime();
    -229/**
    -230 * How many blocks to look at before 
    we give up.
    -231 * There could be many millions of 
    blocks. We don't want the
    -232 * ui to freeze while we run through 
    1B blocks... users will
    -233 * think hbase dead. UI displays 
    warning in red when stats
    -234 * are incomplete.
    -235 */
    -236private final int max;
    -237public static final int DEFAULT_MAX = 
    100;
    -238
    -239CachedBlocksByFile() {
    -240  this(null);
    -241}
    -242
    -243CachedBlocksByFile(final 
    Configuration c) {
    -244  this.max = c == null? DEFAULT_MAX: 
    c.getInt("hbase.ui.blockcache.by.file.max", DEFAULT_MAX);
    -245}
    -246
    -247/**
    -248 * Map by filename. use concurent 
    utils because we want our Map and contained blocks sorted.
    -249 */
    -250private NavigableMapString, 
    NavigableSetCachedBlock cachedBlockByFile = new 
    ConcurrentSkipListMap();
    -251FastLongHistogram hist = new 
    FastLongHistogram();
    -252
    -253/**
    -254 * @param cb
    -255 * @return True if full if we 
    won't be adding any more.
    -256 */
    -257public boolean update(final 
    CachedBlock cb) {
    -258  if (isFull()) return true;
    -259  NavigableSetCachedBlock set 
    = this.cachedBlockByFile.get(cb.getFilename());
    -260  if (set == null) {
    -261set = new 
    ConcurrentSkipListSet();
    -262
    this.cachedBlockByFile.put(cb.getFilename(), set);
    -263  }
    -264  set.add(cb);
    -265  this.size += cb.getSize();
    -266  this.count++;
    -267  BlockType bt = cb.getBlockType();
    -268  if (bt != null  
    bt.isData()) {
    -269this.dataBlockCount++;
    -270this.dataSize += cb.getSize();
    -271  }
    -272  long age = (this.now - 
    cb.getCachedTime())/NANOS_PER_SECOND;
    -273  this.hist.add(age, 1);
    -274  return false;
    -275}
    -276
    -277/**
    -278 * @return True if full; i.e. there 
    are more items in the cache but we only loaded up
    -279 * the maximum set in configuration 
    codehbase.ui.blockcache.by.file.max/code
    -280 * (Default: DEFAULT_MAX).
    -281 */
    -282public boolean isFull() {
    -283  return this.count = this.max;
    -284}
    -285
    -286public NavigableMapString, 
    NavigableSetCachedBlock getCachedBlockStatsByFile() {
    -287  return this.cachedBlockByFile;
    -288}
    -289
    -290/**
    -291 * @return count of blocks in the 
    cache
    -292 */
    -293public int getCount() {
    -294  return count;
    -295}
    -296
    -297public int getDataCount() {
    -298  return dataBlockCount;
    -299}
    -300
    -301/**
    -302 * @return size of blocks in the 
    cache
    -303 */
    -304public long getSize() {
    -305  return size;
    -306}
    -307
    -308/**
    -309 * @return Size of data.
    -310 */
    -311public long getDataSize() {
    -312  return dataSize;
    -313}
    -314
    -315

    [07/51] [partial] hbase-site git commit: Published site at 85b41f36e01214b6485c9352875c84ebf877dab3.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
    index c10cfbf..a3e2f4a 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
    @@ -3371,7 +3371,7 @@
     3363private V result = null;
     3364
     3365private final HBaseAdmin admin;
    -3366private final Long procId;
    +3366protected final Long procId;
     3367
     3368public ProcedureFuture(final 
    HBaseAdmin admin, final Long procId) {
     3369  this.admin = admin;
    @@ -3653,653 +3653,651 @@
     3645 * @return a description of the 
    operation
     3646 */
     3647protected String getDescription() 
    {
    -3648  return "Operation: " + 
    getOperationType() + ", "
    -3649  + "Table Name: " + 
    tableName.getNameWithNamespaceInclAsString();
    -3650
    -3651}
    -3652
    -3653protected abstract class 
    TableWaitForStateCallable implements WaitForStateCallable {
    -3654  @Override
    -3655  public void 
    throwInterruptedException() throws InterruptedIOException {
    -3656throw new 
    InterruptedIOException("Interrupted while waiting for operation: "
    -3657+ getOperationType() + " on 
    table: " + tableName.getNameWithNamespaceInclAsString());
    -3658  }
    -3659
    -3660  @Override
    -3661  public void 
    throwTimeoutException(long elapsedTime) throws TimeoutException {
    -3662throw new TimeoutException("The 
    operation: " + getOperationType() + " on table: " +
    -3663tableName.getNameAsString() 
    + " has not completed after " + elapsedTime + "ms");
    -3664  }
    -3665}
    -3666
    -3667@Override
    -3668protected V 
    postOperationResult(final V result, final long deadlineTs)
    -3669throws IOException, 
    TimeoutException {
    -3670  LOG.info(getDescription() + " 
    completed");
    -3671  return 
    super.postOperationResult(result, deadlineTs);
    -3672}
    -3673
    -3674@Override
    -3675protected V 
    postOperationFailure(final IOException exception, final long deadlineTs)
    -3676throws IOException, 
    TimeoutException {
    -3677  LOG.info(getDescription() + " 
    failed with " + exception.getMessage());
    -3678  return 
    super.postOperationFailure(exception, deadlineTs);
    -3679}
    -3680
    -3681protected void 
    waitForTableEnabled(final long deadlineTs)
    -3682throws IOException, 
    TimeoutException {
    -3683  waitForState(deadlineTs, new 
    TableWaitForStateCallable() {
    -3684@Override
    -3685public boolean checkState(int 
    tries) throws IOException {
    -3686  try {
    -3687if 
    (getAdmin().isTableAvailable(tableName)) {
    -3688  return true;
    -3689}
    -3690  } catch 
    (TableNotFoundException tnfe) {
    -3691LOG.debug("Table " + 
    tableName.getNameWithNamespaceInclAsString()
    -3692+ " was not enabled, 
    sleeping. tries=" + tries);
    -3693  }
    -3694  return false;
    -3695}
    -3696  });
    -3697}
    -3698
    -3699protected void 
    waitForTableDisabled(final long deadlineTs)
    -3700throws IOException, 
    TimeoutException {
    -3701  waitForState(deadlineTs, new 
    TableWaitForStateCallable() {
    -3702@Override
    -3703public boolean checkState(int 
    tries) throws IOException {
    -3704  return 
    getAdmin().isTableDisabled(tableName);
    -3705}
    -3706  });
    -3707}
    -3708
    -3709protected void 
    waitTableNotFound(final long deadlineTs)
    -3710throws IOException, 
    TimeoutException {
    -3711  waitForState(deadlineTs, new 
    TableWaitForStateCallable() {
    -3712@Override
    -3713public boolean checkState(int 
    tries) throws IOException {
    -3714  return 
    !getAdmin().tableExists(tableName);
    -3715}
    -3716  });
    -3717}
    -3718
    -3719protected void 
    waitForSchemaUpdate(final long deadlineTs)
    -3720throws IOException, 
    TimeoutException {
    -3721  waitForState(deadlineTs, new 
    TableWaitForStateCallable() {
    -3722@Override
    -3723public boolean checkState(int 
    tries) throws IOException {
    -3724  return 
    getAdmin().getAlterStatus(tableName).getFirst() == 0;
    -3725}
    -3726  });
    -3727}
    -3728
    -3729protected void 
    waitForAllRegionsOnline(final long deadlineTs, final byte[][] splitKeys)
    -3730throws IOException, 
    TimeoutException {
    -3731  final TableDescriptor desc = 
    getTableDescriptor();
    -3732  final AtomicInteger actualRegCount 
    = new AtomicInteger(0);
    -3733  final MetaTableAccessor.Visitor 
    visitor = new MetaTableAccessor.Visitor() {
    -3734@Override
    -3735public boolean visit(Result 
    rowResult) throws IOException {
    -3736  

    [07/51] [partial] hbase-site git commit: Published site at 6198e1fc7dfa85c3bc6b2855f9a5fb5f4b2354ff.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
    index cc07d09..5db3a2e 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
    @@ -101,7 +101,7 @@ var activeTableTab = "activeTableTab";
     
     
     All Known Implementing Classes:
    -AbstractPeerProcedure, AddPeerProcedure, DisablePeerProcedure, EnablePeerProcedure, ModifyPeerProcedure, RefreshPeerProcedure, RemovePeerProcedure, UpdatePeerConfigProcedure
    +AbstractPeerProcedure, AddPeerProcedure, DisablePeerProcedure, EnablePeerProcedure, ModifyPeerProcedure, RecoverStandbyProcedure, RefreshPeerProcedure, RemovePeerProcedure, SyncReplicationReplayWALProcedure,
     SyncReplicationReplayWALRemoteProcedure,
     TransitPeerSyncReplicationStateProcedure,
     UpdatePeerConfigProcedure
     
     
     
    @@ -173,7 +173,7 @@ public interface 
     
     getPeerId
    -https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">StringgetPeerId()
    +https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">StringgetPeerId()
     
     
     
    @@ -182,7 +182,7 @@ public interface 
     
     getPeerOperationType
    -PeerProcedureInterface.PeerOperationTypegetPeerOperationType()
    +PeerProcedureInterface.PeerOperationTypegetPeerOperationType()
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
    index 792aeaf..f98ae5c 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
    @@ -2093,57 +2093,93 @@
     RefreshPeerProcedure.abort(MasterProcedureEnvenv)
     
     
    +protected boolean
    +SyncReplicationReplayWALRemoteProcedure.abort(MasterProcedureEnvenv)
    +
    +
     protected Procedure.LockState
     AbstractPeerProcedure.acquireLock(MasterProcedureEnvenv)
     
    -
    +
     private void
     RefreshPeerProcedure.complete(MasterProcedureEnvenv,
     https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
     title="class or interface in 
    java.lang">Throwableerror)
     
    +
    +private void
    +SyncReplicationReplayWALRemoteProcedure.complete(MasterProcedureEnvenv,
    +https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
     title="class or interface in 
    java.lang">Throwableerror)
    +
     
    +private void
    +TransitPeerSyncReplicationStateProcedure.createDirForRemoteWAL(MasterProcedureEnvenv)
    +
    +
     protected ProcedureMasterProcedureEnv[]
     RefreshPeerProcedure.execute(MasterProcedureEnvenv)
     
    +
    +protected ProcedureMasterProcedureEnv[]
    +SyncReplicationReplayWALRemoteProcedure.execute(MasterProcedureEnvenv)
    +
     
     protected StateMachineProcedure.Flow
     ModifyPeerProcedure.executeFromState(MasterProcedureEnvenv,
     
    org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.PeerModificationStatestate)
     
     
    +protected StateMachineProcedure.Flow
    +TransitPeerSyncReplicationStateProcedure.executeFromState(MasterProcedureEnvenv,
    +
    org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.PeerSyncReplicationStateTransitionStatestate)
    +
    +
    +protected StateMachineProcedure.Flow
    +RecoverStandbyProcedure.executeFromState(MasterProcedureEnvenv,
    +
    org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RecoverStandbyStatestate)
    +
    +
    +protected StateMachineProcedure.Flow
    +SyncReplicationReplayWALProcedure.executeFromState(MasterProcedureEnvenv,
    +
    org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SyncReplicationReplayWALStatestate)
    +
    +
     protected boolean
     AbstractPeerProcedure.hasLock(MasterProcedureEnvenv)
     
    -
    +
     protected boolean
     AbstractPeerProcedure.holdLock(MasterProcedureEnvenv)
     
    -
    +
     protected void
     AddPeerProcedure.postPeerModification(MasterProcedureEnvenv)
     
    -
    +
     protected void
     EnablePeerProcedure.postPeerModification(MasterProcedureEnvenv)
     
    -
    +
     protected void
     UpdatePeerConfigProcedure.postPeerModification(MasterProcedureEnvenv)
     
    -
    +
     protected abstract void
     ModifyPeerProcedure.postPeerModification(MasterProcedureEnvenv)
     Called before we finish the procedure.
     
     
    -
    +
     protected void
     

    [07/51] [partial] hbase-site git commit: Published site at 14087cc919da9f2e0b1a68f701f6365ad9d1d71f.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html 
    b/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html
    index de1962a..db0fab9 100644
    --- a/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html
    +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html
    @@ -18,7 +18,7 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":6,"i1":6,"i2":18,"i3":6,"i4":18,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":38,"i46":38,"i47":38,"i48":6,"i49":6,"i50":6,"i51":6,"i52":6,"i53":6,"i54":6,"i55":6};
    +var methods = 
    {"i0":6,"i1":6,"i2":18,"i3":6,"i4":18,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":38,"i47":38,"i48":38,"i49":6,"i50":6,"i51":6,"i52":6,"i53":6,"i54":6,"i55":6,"i56":6};
     var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
    Methods"],4:["t3","Abstract Methods"],16:["t5","Default 
    Methods"],32:["t6","Deprecated Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
    @@ -278,69 +278,73 @@ extends 
     long
    -getDataInMemoryWithoutWAL()
    +getCpRequestsCount()
     
     
     long
    -getEarliestFlushTimeForAllStores()
    +getDataInMemoryWithoutWAL()
     
     
     long
    -getFilteredReadRequestsCount()
    +getEarliestFlushTimeForAllStores()
     
     
     long
    -getMaxFlushedSeqId()
    +getFilteredReadRequestsCount()
     
     
    +long
    +getMaxFlushedSeqId()
    +
    +
     https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
     title="class or interface in java.util">Mapbyte[],https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
     title="class or interface in java.lang">Long
     getMaxStoreSeqId()
     
    -
    +
     long
     getMemStoreDataSize()
     
    -
    +
     long
     getMemStoreHeapSize()
     
    -
    +
     long
     getMemStoreOffHeapSize()
     
    -
    +
     long
     getNumMutationsWithoutWAL()
     
    -
    +
     long
     getOldestHfileTs(booleanmajorCompactionOnly)
     This can be used to determine the last time all files of 
    this region were major compacted.
     
     
    -
    +
     long
     getReadRequestsCount()
     
    -
    +
     RegionInfo
     getRegionInfo()
     
    -
    +
     Region.RowLock
     getRowLock(byte[]row,
       booleanreadLock)
     Get a row lock for the specified row.
     
     
    -
    +
     RegionScanner
     getScanner(Scanscan)
     Return an iterator that scans over the HRegion, returning 
    the indicated
      columns and rows specified by the Scan.
     
     
    -
    +
     RegionScanner
     getScanner(Scanscan,
       https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
     title="class or interface in java.util">ListKeyValueScanneradditionalScanners)
    @@ -348,67 +352,67 @@ extends Scan.
     
     
    -
    +
     Store
     getStore(byte[]family)
     Return the Store for the given family
     
     
    -
    +
     https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
     title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
     getStoreFileList(byte[][]columns)
     
    -
    +
     https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
     title="class or interface in java.util">List? extends Store
     getStores()
     Return the list of Stores managed by this region
     
     
    -
    +
     TableDescriptor
     getTableDescriptor()
     
    -
    +
     long
     getWriteRequestsCount()
     
    -
    +
     Result
     increment(Incrementincrement)
     Perform one or more increment operations on a row.
     
     
    -
    +
     boolean
     isAvailable()
     
    -
    +
     boolean
     isClosed()
     
    -
    +
     boolean
     isClosing()
     
    -
    +
     boolean
     isMergeable()
     
    -
    +
     boolean
     isReadOnly()
     
    -
    +
     boolean
     isSplittable()
     
    -
    +
     void
     mutateRow(RowMutationsmutations)
     Performs multiple mutations atomically on a single 
    row.
     
     
    -
    +
     void
     mutateRowsWithLocks(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
     title="class or interface in java.util">CollectionMutationmutations,
    https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
     title="class or interface in 
    java.util">Collectionbyte[]rowsToLock,
    @@ -417,7 +421,7 @@ extends Perform atomic mutations within the region.
     
     
    -
    +
     void
     processRowsWithLocks(RowProcessor?,?processor)
     Deprecated.
    @@ -426,7 +430,7 @@ extends 
    +
     void
     processRowsWithLocks(RowProcessor?,?processor,
     longnonceGroup,
    @@ -437,7 +441,7 @@ extends 
    +
     void
     processRowsWithLocks(RowProcessor?,?processor,
     

    [07/51] [partial] hbase-site git commit: Published site at 72784c2d836a4b977667449d3adec5e8d15453f5.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
    new file mode 100644
    index 000..546d2b6
    --- /dev/null
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
    @@ -0,0 +1,3793 @@
    +http://www.w3.org/TR/html4/loose.dtd;>
    +
    +
    +Source code
    +
    +
    +
    +
    +001/**
    +002 * Licensed to the Apache Software 
    Foundation (ASF) under one
    +003 * or more contributor license 
    agreements.  See the NOTICE file
    +004 * distributed with this work for 
    additional information
    +005 * regarding copyright ownership.  The 
    ASF licenses this file
    +006 * to you under the Apache License, 
    Version 2.0 (the
    +007 * "License"); you may not use this file 
    except in compliance
    +008 * with the License.  You may obtain a 
    copy of the License at
    +009 *
    +010 * 
    http://www.apache.org/licenses/LICENSE-2.0
    +011 *
    +012 * Unless required by applicable law or 
    agreed to in writing, software
    +013 * distributed under the License is 
    distributed on an "AS IS" BASIS,
    +014 * WITHOUT WARRANTIES OR CONDITIONS OF 
    ANY KIND, either express or implied.
    +015 * See the License for the specific 
    language governing permissions and
    +016 * limitations under the License.
    +017 */
    +018package org.apache.hadoop.hbase.master;
    +019
    +020import static 
    org.apache.hadoop.hbase.HConstants.HBASE_MASTER_LOGCLEANER_PLUGINS;
    +021
    +022import com.google.protobuf.Descriptors;
    +023import com.google.protobuf.Service;
    +024import java.io.IOException;
    +025import java.io.InterruptedIOException;
    +026import java.lang.reflect.Constructor;
    +027import 
    java.lang.reflect.InvocationTargetException;
    +028import java.net.InetAddress;
    +029import java.net.InetSocketAddress;
    +030import java.net.UnknownHostException;
    +031import java.util.ArrayList;
    +032import java.util.Arrays;
    +033import java.util.Collection;
    +034import java.util.Collections;
    +035import java.util.Comparator;
    +036import java.util.EnumSet;
    +037import java.util.HashMap;
    +038import java.util.Iterator;
    +039import java.util.List;
    +040import java.util.Map;
    +041import java.util.Map.Entry;
    +042import java.util.Objects;
    +043import java.util.Optional;
    +044import java.util.Set;
    +045import 
    java.util.concurrent.ExecutionException;
    +046import java.util.concurrent.Future;
    +047import java.util.concurrent.TimeUnit;
    +048import 
    java.util.concurrent.TimeoutException;
    +049import 
    java.util.concurrent.atomic.AtomicInteger;
    +050import 
    java.util.concurrent.atomic.AtomicReference;
    +051import java.util.function.Function;
    +052import java.util.regex.Pattern;
    +053import java.util.stream.Collectors;
    +054import javax.servlet.ServletException;
    +055import javax.servlet.http.HttpServlet;
    +056import 
    javax.servlet.http.HttpServletRequest;
    +057import 
    javax.servlet.http.HttpServletResponse;
    +058import 
    org.apache.commons.lang3.StringUtils;
    +059import 
    org.apache.hadoop.conf.Configuration;
    +060import org.apache.hadoop.fs.Path;
    +061import 
    org.apache.hadoop.hbase.ClusterId;
    +062import 
    org.apache.hadoop.hbase.ClusterMetrics;
    +063import 
    org.apache.hadoop.hbase.ClusterMetrics.Option;
    +064import 
    org.apache.hadoop.hbase.ClusterMetricsBuilder;
    +065import 
    org.apache.hadoop.hbase.DoNotRetryIOException;
    +066import 
    org.apache.hadoop.hbase.HBaseIOException;
    +067import 
    org.apache.hadoop.hbase.HBaseInterfaceAudience;
    +068import 
    org.apache.hadoop.hbase.HConstants;
    +069import 
    org.apache.hadoop.hbase.InvalidFamilyOperationException;
    +070import 
    org.apache.hadoop.hbase.MasterNotRunningException;
    +071import 
    org.apache.hadoop.hbase.MetaTableAccessor;
    +072import 
    org.apache.hadoop.hbase.NamespaceDescriptor;
    +073import 
    org.apache.hadoop.hbase.PleaseHoldException;
    +074import 
    org.apache.hadoop.hbase.ReplicationPeerNotFoundException;
    +075import 
    org.apache.hadoop.hbase.ScheduledChore;
    +076import 
    org.apache.hadoop.hbase.ServerName;
    +077import 
    org.apache.hadoop.hbase.TableDescriptors;
    +078import 
    org.apache.hadoop.hbase.TableName;
    +079import 
    org.apache.hadoop.hbase.TableNotDisabledException;
    +080import 
    org.apache.hadoop.hbase.TableNotFoundException;
    +081import 
    org.apache.hadoop.hbase.UnknownRegionException;
    +082import 
    org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
    +083import 
    org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
    +084import 
    org.apache.hadoop.hbase.client.MasterSwitchType;
    +085import 
    org.apache.hadoop.hbase.client.RegionInfo;
    +086import 
    org.apache.hadoop.hbase.client.RegionInfoBuilder;
    +087import 
    org.apache.hadoop.hbase.client.Result;
    +088import 
    org.apache.hadoop.hbase.client.TableDescriptor;
    +089import 
    org.apache.hadoop.hbase.client.TableDescriptorBuilder;
    +090import 
    org.apache.hadoop.hbase.client.TableState;
    +091import 
    

    [07/51] [partial] hbase-site git commit: Published site at 9101fc246f86445006bfbcdfda5cc495016dc280.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.FlushedSequenceIdFlusher.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.FlushedSequenceIdFlusher.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.FlushedSequenceIdFlusher.html
    new file mode 100644
    index 000..ba6e4aa
    --- /dev/null
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.FlushedSequenceIdFlusher.html
    @@ -0,0 +1,1244 @@
    +http://www.w3.org/TR/html4/loose.dtd;>
    +
    +
    +Source code
    +
    +
    +
    +
    +001/*
    +002 *
    +003 * Licensed to the Apache Software 
    Foundation (ASF) under one
    +004 * or more contributor license 
    agreements.  See the NOTICE file
    +005 * distributed with this work for 
    additional information
    +006 * regarding copyright ownership.  The 
    ASF licenses this file
    +007 * to you under the Apache License, 
    Version 2.0 (the
    +008 * "License"); you may not use this file 
    except in compliance
    +009 * with the License.  You may obtain a 
    copy of the License at
    +010 *
    +011 * 
    http://www.apache.org/licenses/LICENSE-2.0
    +012 *
    +013 * Unless required by applicable law or 
    agreed to in writing, software
    +014 * distributed under the License is 
    distributed on an "AS IS" BASIS,
    +015 * WITHOUT WARRANTIES OR CONDITIONS OF 
    ANY KIND, either express or implied.
    +016 * See the License for the specific 
    language governing permissions and
    +017 * limitations under the License.
    +018 */
    +019package org.apache.hadoop.hbase.master;
    +020
    +021import static 
    org.apache.hadoop.hbase.util.CollectionUtils.computeIfAbsent;
    +022
    +023import java.io.IOException;
    +024import java.net.InetAddress;
    +025import java.util.ArrayList;
    +026import java.util.Collections;
    +027import java.util.HashMap;
    +028import java.util.Iterator;
    +029import java.util.List;
    +030import java.util.Map;
    +031import java.util.Map.Entry;
    +032import java.util.Set;
    +033import 
    java.util.concurrent.ConcurrentNavigableMap;
    +034import 
    java.util.concurrent.ConcurrentSkipListMap;
    +035import 
    java.util.concurrent.CopyOnWriteArrayList;
    +036import 
    java.util.concurrent.atomic.AtomicBoolean;
    +037import java.util.function.Predicate;
    +038import 
    org.apache.hadoop.conf.Configuration;
    +039import 
    org.apache.hadoop.fs.FSDataInputStream;
    +040import 
    org.apache.hadoop.fs.FSDataOutputStream;
    +041import org.apache.hadoop.fs.FileSystem;
    +042import org.apache.hadoop.fs.Path;
    +043import 
    org.apache.hadoop.hbase.ClockOutOfSyncException;
    +044import 
    org.apache.hadoop.hbase.HConstants;
    +045import 
    org.apache.hadoop.hbase.NotServingRegionException;
    +046import 
    org.apache.hadoop.hbase.RegionMetrics;
    +047import 
    org.apache.hadoop.hbase.ScheduledChore;
    +048import 
    org.apache.hadoop.hbase.ServerMetrics;
    +049import 
    org.apache.hadoop.hbase.ServerMetricsBuilder;
    +050import 
    org.apache.hadoop.hbase.ServerName;
    +051import 
    org.apache.hadoop.hbase.YouAreDeadException;
    +052import 
    org.apache.hadoop.hbase.client.ClusterConnection;
    +053import 
    org.apache.hadoop.hbase.client.RegionInfo;
    +054import 
    org.apache.hadoop.hbase.client.RetriesExhaustedException;
    +055import 
    org.apache.hadoop.hbase.ipc.HBaseRpcController;
    +056import 
    org.apache.hadoop.hbase.ipc.RpcControllerFactory;
    +057import 
    org.apache.hadoop.hbase.master.assignment.RegionStates;
    +058import 
    org.apache.hadoop.hbase.monitoring.MonitoredTask;
    +059import 
    org.apache.hadoop.hbase.regionserver.HRegionServer;
    +060import 
    org.apache.hadoop.hbase.util.Bytes;
    +061import 
    org.apache.hadoop.hbase.util.FSUtils;
    +062import 
    org.apache.hadoop.hbase.zookeeper.ZKUtil;
    +063import 
    org.apache.hadoop.hbase.zookeeper.ZKWatcher;
    +064import 
    org.apache.yetus.audience.InterfaceAudience;
    +065import 
    org.apache.zookeeper.KeeperException;
    +066import org.slf4j.Logger;
    +067import org.slf4j.LoggerFactory;
    +068
    +069import 
    org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
    +070import 
    org.apache.hbase.thirdparty.com.google.protobuf.ByteString;
    +071import 
    org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations;
    +072
    +073import 
    org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
    +074import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
    +075import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds;
    +076import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId;
    +077import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.FlushedRegionSequenceId;
    +078import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.FlushedSequenceId;
    +079import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.FlushedStoreSequenceId;
    +080import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest;
    +081
    +082/**
    +083 * The ServerManager class manages info 
    about region servers.
    +084 * p
    +085 

    [07/51] [partial] hbase-site git commit: Published site at 0b28155d274910b4e667b949d51f78809a1eff0b.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.StatelessUncompressor.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.StatelessUncompressor.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.StatelessUncompressor.html
    new file mode 100644
    index 000..9df0225
    --- /dev/null
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.StatelessUncompressor.html
    @@ -0,0 +1,476 @@
    +http://www.w3.org/TR/html4/loose.dtd;>
    +
    +
    +Source code
    +
    +
    +
    +
    +001/*
    +002 * Licensed to the Apache Software 
    Foundation (ASF) under one
    +003 * or more contributor license 
    agreements.  See the NOTICE file
    +004 * distributed with this work for 
    additional information
    +005 * regarding copyright ownership.  The 
    ASF licenses this file
    +006 * to you under the Apache License, 
    Version 2.0 (the
    +007 * "License"); you may not use this file 
    except in compliance
    +008 * with the License.  You may obtain a 
    copy of the License at
    +009 *
    +010 * 
    http://www.apache.org/licenses/LICENSE-2.0
    +011 *
    +012 * Unless required by applicable law or 
    agreed to in writing, software
    +013 * distributed under the License is 
    distributed on an "AS IS" BASIS,
    +014 * WITHOUT WARRANTIES OR CONDITIONS OF 
    ANY KIND, either express or implied.
    +015 * See the License for the specific 
    language governing permissions and
    +016 * limitations under the License.
    +017 */
    +018package 
    org.apache.hadoop.hbase.regionserver.wal;
    +019
    +020import java.io.ByteArrayOutputStream;
    +021import java.io.IOException;
    +022import java.io.InputStream;
    +023import java.io.OutputStream;
    +024
    +025import 
    org.apache.hadoop.conf.Configuration;
    +026import org.apache.hadoop.hbase.Cell;
    +027import 
    org.apache.hadoop.hbase.CellUtil;
    +028import 
    org.apache.hadoop.hbase.HBaseInterfaceAudience;
    +029import 
    org.apache.hadoop.hbase.PrivateCellUtil;
    +030import 
    org.apache.hadoop.hbase.KeyValue;
    +031import 
    org.apache.hadoop.hbase.KeyValueUtil;
    +032import 
    org.apache.yetus.audience.InterfaceAudience;
    +033import 
    org.apache.hadoop.hbase.codec.BaseDecoder;
    +034import 
    org.apache.hadoop.hbase.codec.BaseEncoder;
    +035import 
    org.apache.hadoop.hbase.codec.Codec;
    +036import 
    org.apache.hadoop.hbase.codec.KeyValueCodecWithTags;
    +037import 
    org.apache.hadoop.hbase.io.ByteBuffInputStream;
    +038import 
    org.apache.hadoop.hbase.io.ByteBufferWriter;
    +039import 
    org.apache.hadoop.hbase.io.ByteBufferWriterOutputStream;
    +040import 
    org.apache.hadoop.hbase.io.util.Dictionary;
    +041import 
    org.apache.hadoop.hbase.io.util.StreamUtils;
    +042import 
    org.apache.hadoop.hbase.nio.ByteBuff;
    +043import 
    org.apache.hadoop.hbase.util.ByteBufferUtils;
    +044import 
    org.apache.hadoop.hbase.util.Bytes;
    +045import 
    org.apache.hadoop.hbase.util.ReflectionUtils;
    +046import org.apache.hadoop.io.IOUtils;
    +047
    +048import 
    org.apache.hbase.thirdparty.com.google.protobuf.ByteString;
    +049import 
    org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations;
    +050
    +051
    +052/**
    +053 * Compression in this class is lifted 
    off Compressor/KeyValueCompression.
    +054 * This is a pure coincidence... they are 
    independent and don't have to be compatible.
    +055 *
    +056 * This codec is used at server side for 
    writing cells to WAL as well as for sending edits
    +057 * as part of the distributed splitting 
    process.
    +058 */
    +059@InterfaceAudience.LimitedPrivate({HBaseInterfaceAudience.COPROC,
    +060  HBaseInterfaceAudience.PHOENIX, 
    HBaseInterfaceAudience.CONFIG})
    +061public class WALCellCodec implements 
    Codec {
    +062  /** Configuration key for the class to 
    use when encoding cells in the WAL */
    +063  public static final String 
    WAL_CELL_CODEC_CLASS_KEY = "hbase.regionserver.wal.codec";
    +064
    +065  protected final CompressionContext 
    compression;
    +066
    +067  /**
    +068   * bAll subclasses must 
    implement a no argument constructor/b
    +069   */
    +070  public WALCellCodec() {
    +071this.compression = null;
    +072  }
    +073
    +074  /**
    +075   * Default constructor - ball 
    subclasses must implement a constructor with this signature /b
    +076   * if they are to be dynamically loaded 
    from the {@link Configuration}.
    +077   * @param conf configuration to 
    configure ttthis/tt
    +078   * @param compression compression the 
    codec should support, can be ttnull/tt to indicate no
    +079   *  compression
    +080   */
    +081  public WALCellCodec(Configuration conf, 
    CompressionContext compression) {
    +082this.compression = compression;
    +083  }
    +084
    +085  public static String 
    getWALCellCodecClass(Configuration conf) {
    +086return 
    conf.get(WAL_CELL_CODEC_CLASS_KEY, WALCellCodec.class.getName());
    +087  }
    +088
    +089  /**
    +090   * Create and setup a {@link 
    WALCellCodec} from the {@code cellCodecClsName} and
    +091   * CompressionContext, if {@code 
    cellCodecClsName} is specified.
    +092   * Otherwise Cell Codec classname is 
    read from 

    [07/51] [partial] hbase-site git commit: Published site at 7d3750bd9fc9747623549c242cc4171e224b3eaf.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
    index 594ef24..17d5c40 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
    @@ -170,241 +170,242 @@
     162  }
     163
     164  /**
    -165   * Add a remote rpc. Be sure to check 
    result for successful add.
    +165   * Add a remote rpc.
     166   * @param key the node identifier
    -167   * @return True if we successfully 
    added the operation.
    -168   */
    -169  public boolean addOperationToNode(final 
    TRemote key, RemoteProcedure rp) {
    +167   */
    +168  public void addOperationToNode(final 
    TRemote key, RemoteProcedure rp)
    +169  throws 
    NullTargetServerDispatchException, NoServerDispatchException, 
    NoNodeDispatchException {
     170if (key == null) {
    -171  // Key is remote server name. Be 
    careful. It could have been nulled by a concurrent
    -172  // ServerCrashProcedure shutting 
    down outstanding RPC requests. See remoteCallFailed.
    -173  return false;
    -174}
    -175assert key != null : "found null key 
    for node";
    -176BufferNode node = nodeMap.get(key);
    -177if (node == null) {
    -178  return false;
    -179}
    -180node.add(rp);
    -181// Check our node still in the map; 
    could have been removed by #removeNode.
    -182return nodeMap.containsValue(node);
    -183  }
    -184
    -185  /**
    -186   * Remove a remote node
    -187   * @param key the node identifier
    -188   */
    -189  public boolean removeNode(final TRemote 
    key) {
    -190final BufferNode node = 
    nodeMap.remove(key);
    -191if (node == null) return false;
    -192node.abortOperationsInQueue();
    -193return true;
    -194  }
    -195
    -196  // 
    
    -197  //  Task Helpers
    -198  // 
    
    -199  protected FutureVoid 
    submitTask(CallableVoid task) {
    -200return threadPool.submit(task);
    -201  }
    -202
    -203  protected FutureVoid 
    submitTask(CallableVoid task, long delay, TimeUnit unit) {
    -204final FutureTaskVoid 
    futureTask = new FutureTask(task);
    -205timeoutExecutor.add(new 
    DelayedTask(futureTask, delay, unit));
    -206return futureTask;
    -207  }
    -208
    -209  protected abstract void 
    remoteDispatch(TRemote key, SetRemoteProcedure operations);
    -210  protected abstract void 
    abortPendingOperations(TRemote key, SetRemoteProcedure operations);
    -211
    -212  /**
    -213   * Data structure with reference to 
    remote operation.
    -214   */
    -215  public static abstract class 
    RemoteOperation {
    -216private final RemoteProcedure 
    remoteProcedure;
    -217
    -218protected RemoteOperation(final 
    RemoteProcedure remoteProcedure) {
    -219  this.remoteProcedure = 
    remoteProcedure;
    -220}
    -221
    -222public RemoteProcedure 
    getRemoteProcedure() {
    -223  return remoteProcedure;
    -224}
    -225  }
    -226
    -227  /**
    -228   * Remote procedure reference.
    -229   */
    -230  public interface 
    RemoteProcedureTEnv, TRemote {
    -231/**
    -232 * For building the remote 
    operation.
    -233 */
    -234RemoteOperation remoteCallBuild(TEnv 
    env, TRemote remote);
    -235
    -236/**
    -237 * Called when the executeProcedure 
    call is failed.
    -238 */
    -239void remoteCallFailed(TEnv env, 
    TRemote remote, IOException exception);
    -240
    -241/**
    -242 * Called when RS tells the remote 
    procedure is succeeded through the
    -243 * {@code reportProcedureDone} 
    method.
    -244 */
    -245void remoteOperationCompleted(TEnv 
    env);
    -246
    -247/**
    -248 * Called when RS tells the remote 
    procedure is failed through the {@code reportProcedureDone}
    -249 * method.
    -250 */
    -251void remoteOperationFailed(TEnv env, 
    RemoteProcedureException error);
    -252  }
    -253
    -254  /**
    -255   * Account of what procedures are 
    running on remote node.
    -256   * @param TEnv
    -257   * @param TRemote
    -258   */
    -259  public interface RemoteNodeTEnv, 
    TRemote {
    -260TRemote getKey();
    -261void add(RemoteProcedureTEnv, 
    TRemote operation);
    -262void dispatch();
    -263  }
    -264
    -265  protected 
    ArrayListMultimapClass?, RemoteOperation 
    buildAndGroupRequestByType(final TEnv env,
    -266  final TRemote remote, final 
    SetRemoteProcedure remoteProcedures) {
    -267final 
    ArrayListMultimapClass?, RemoteOperation requestByType = 
    ArrayListMultimap.create();
    -268for (RemoteProcedure proc: 
    remoteProcedures) {
    -269  RemoteOperation operation = 
    proc.remoteCallBuild(env, remote);
    -270  
    requestByType.put(operation.getClass(), 

    [07/51] [partial] hbase-site git commit: Published site at 997747076d8ec0b4346d7cb99c4b0667a7c14905.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
    --
    diff --git 
    a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html 
    b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
    index d2f8d9e..43fd3a1 100644
    --- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
    +++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
    @@ -947,674 +947,678 @@
     
     
     static HBaseClassTestRule
    -TestImmutableHTableDescriptor.CLASS_RULE
    +TestTableDescriptorUtils.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestTableDescriptorBuilder.CLASS_RULE
    +TestImmutableHTableDescriptor.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestBufferedMutatorParams.CLASS_RULE
    +TestTableDescriptorBuilder.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncProcess.CLASS_RULE
    +TestBufferedMutatorParams.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestClientScanner.CLASS_RULE
    +TestAsyncProcess.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestClientNoCluster.CLASS_RULE
    +TestClientScanner.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAttributes.CLASS_RULE
    +TestClientNoCluster.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestMetricsConnection.CLASS_RULE
    +TestAttributes.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestSnapshotFromAdmin.CLASS_RULE
    +TestMetricsConnection.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestDeleteTimeStamp.CLASS_RULE
    +TestSnapshotFromAdmin.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncProcessWithRegionException.CLASS_RULE
    +TestDeleteTimeStamp.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestHTableMultiplexerViaMocks.CLASS_RULE
    +TestAsyncProcessWithRegionException.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestImmutableHRegionInfo.CLASS_RULE
    +TestHTableMultiplexerViaMocks.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestDelayingRunner.CLASS_RULE
    +TestImmutableHRegionInfo.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestReversedScannerCallable.CLASS_RULE
    +TestDelayingRunner.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestCoprocessorDescriptor.CLASS_RULE
    +TestReversedScannerCallable.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestClientExponentialBackoff.CLASS_RULE
    +TestCoprocessorDescriptor.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestInterfaceAlign.CLASS_RULE
    +TestClientExponentialBackoff.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestIncrement.CLASS_RULE
    +TestInterfaceAlign.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestColumnFamilyDescriptorBuilder.CLASS_RULE
    +TestIncrement.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestOperation.CLASS_RULE
    +TestColumnFamilyDescriptorBuilder.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestMutation.CLASS_RULE
    +TestOperation.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestScan.CLASS_RULE
    +TestMutation.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestRetriesExhaustedWithDetailsException.CLASS_RULE
    +TestScan.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestSimpleRequestController.CLASS_RULE
    +TestRetriesExhaustedWithDetailsException.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestRowComparator.CLASS_RULE
    +TestSimpleRequestController.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestProcedureFuture.CLASS_RULE
    +TestRowComparator.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestRegionInfoDisplay.CLASS_RULE
    +TestProcedureFuture.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestPutDotHas.CLASS_RULE
    +TestRegionInfoDisplay.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestGet.CLASS_RULE
    +TestPutDotHas.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestImmutableHColumnDescriptor.CLASS_RULE
    +TestGet.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestBufferedMutator.CLASS_RULE
    +TestImmutableHColumnDescriptor.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncNamespaceAdminApi.CLASS_RULE
    +TestBufferedMutator.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestRawAsyncScanCursor.CLASS_RULE
    +TestAsyncNamespaceAdminApi.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestReplicasClient.CLASS_RULE
    +TestRawAsyncScanCursor.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestGetProcedureResult.CLASS_RULE
    +TestReplicasClient.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncReplicationAdminApi.CLASS_RULE
    +TestGetProcedureResult.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncSnapshotAdminApi.CLASS_RULE
    +TestAsyncReplicationAdminApi.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestClientTimeouts.CLASS_RULE
    +TestAsyncSnapshotAdminApi.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncTable.CLASS_RULE
    +TestClientTimeouts.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestCIPutRpcTimeout.CLASS_RULE
    +TestAsyncTable.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestRestoreSnapshotFromClientWithRegionReplicas.CLASS_RULE
    +TestCIPutRpcTimeout.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestLimitedScanWithFilter.CLASS_RULE
    +TestRestoreSnapshotFromClientWithRegionReplicas.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestDropTimeoutRequest.CLASS_RULE
    

    [07/51] [partial] hbase-site git commit: Published site at f3d1c021de2264301f68eadb9ef126ff83d7ef53.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html 
    b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
    index 28539e5..c3f794b 100644
    --- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
    +++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
    @@ -55,404 +55,406 @@
     047 * implementations:
     048 * ul
     049 *   
    liemdefaultProvider/em : whatever provider is standard 
    for the hbase version. Currently
    -050 *  
    "filesystem"/li
    -051 *   
    liemfilesystem/em : a provider that will run on top of 
    an implementation of the Hadoop
    -052 * FileSystem 
    interface, normally HDFS./li
    -053 *   
    liemmultiwal/em : a provider that will use multiple 
    "filesystem" wal instances per region
    -054 *   
    server./li
    -055 * /ul
    -056 *
    -057 * Alternatively, you may provide a 
    custom implementation of {@link WALProvider} by class name.
    -058 */
    -059@InterfaceAudience.Private
    -060public class WALFactory {
    -061
    -062  private static final Logger LOG = 
    LoggerFactory.getLogger(WALFactory.class);
    +050 *  
    "asyncfs"/li
    +051 *   
    liemasyncfs/em : a provider that will run on top of an 
    implementation of the Hadoop
    +052 * FileSystem 
    interface via an asynchronous client./li
    +053 *   
    liemfilesystem/em : a provider that will run on top of 
    an implementation of the Hadoop
    +054 * FileSystem 
    interface via HDFS's synchronous DFSClient./li
    +055 *   
    liemmultiwal/em : a provider that will use multiple 
    "filesystem" wal instances per region
    +056 *   
    server./li
    +057 * /ul
    +058 *
    +059 * Alternatively, you may provide a 
    custom implementation of {@link WALProvider} by class name.
    +060 */
    +061@InterfaceAudience.Private
    +062public class WALFactory {
     063
    -064  /**
    -065   * Maps between configuration names for 
    providers and implementation classes.
    -066   */
    -067  static enum Providers {
    -068
    defaultProvider(AsyncFSWALProvider.class),
    -069filesystem(FSHLogProvider.class),
    -070
    multiwal(RegionGroupingProvider.class),
    -071asyncfs(AsyncFSWALProvider.class);
    -072
    -073final Class? extends 
    WALProvider clazz;
    -074Providers(Class? extends 
    WALProvider clazz) {
    -075  this.clazz = clazz;
    -076}
    -077  }
    -078
    -079  public static final String WAL_PROVIDER 
    = "hbase.wal.provider";
    -080  static final String 
    DEFAULT_WAL_PROVIDER = Providers.defaultProvider.name();
    -081
    -082  public static final String 
    META_WAL_PROVIDER = "hbase.wal.meta_provider";
    -083  static final String 
    DEFAULT_META_WAL_PROVIDER = Providers.defaultProvider.name();
    -084
    -085  final String factoryId;
    -086  private final WALProvider provider;
    -087  // The meta updates are written to a 
    different wal. If this
    -088  // regionserver holds meta regions, 
    then this ref will be non-null.
    -089  // lazily intialized; most 
    RegionServers don't deal with META
    -090  private final 
    AtomicReferenceWALProvider metaProvider = new 
    AtomicReference();
    -091
    -092  /**
    -093   * Configuration-specified WAL Reader 
    used when a custom reader is requested
    -094   */
    -095  private final Class? extends 
    AbstractFSWALProvider.Reader logReaderClass;
    -096
    -097  /**
    -098   * How long to attempt opening 
    in-recovery wals
    -099   */
    -100  private final int timeoutMillis;
    -101
    -102  private final Configuration conf;
    +064  private static final Logger LOG = 
    LoggerFactory.getLogger(WALFactory.class);
    +065
    +066  /**
    +067   * Maps between configuration names for 
    providers and implementation classes.
    +068   */
    +069  static enum Providers {
    +070
    defaultProvider(AsyncFSWALProvider.class),
    +071filesystem(FSHLogProvider.class),
    +072
    multiwal(RegionGroupingProvider.class),
    +073asyncfs(AsyncFSWALProvider.class);
    +074
    +075final Class? extends 
    WALProvider clazz;
    +076Providers(Class? extends 
    WALProvider clazz) {
    +077  this.clazz = clazz;
    +078}
    +079  }
    +080
    +081  public static final String WAL_PROVIDER 
    = "hbase.wal.provider";
    +082  static final String 
    DEFAULT_WAL_PROVIDER = Providers.defaultProvider.name();
    +083
    +084  public static final String 
    META_WAL_PROVIDER = "hbase.wal.meta_provider";
    +085  static final String 
    DEFAULT_META_WAL_PROVIDER = Providers.defaultProvider.name();
    +086
    +087  final String factoryId;
    +088  private final WALProvider provider;
    +089  // The meta updates are written to a 
    different wal. If this
    +090  // regionserver holds meta regions, 
    then this ref will be non-null.
    +091  // lazily intialized; most 
    RegionServers don't deal with META
    +092  private final 
    AtomicReferenceWALProvider metaProvider = new 
    AtomicReference();
    +093
    +094  /**
    +095   * 

    [07/51] [partial] hbase-site git commit: Published site at cf529f18a9959589fa635f78df4840472526ea2c.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
    --
    diff --git 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
     
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
    index 3f8844b..cdb9398 100644
    --- 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
    +++ 
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
    @@ -140,2712 +140,2713 @@
     132public class PerformanceEvaluation 
    extends Configured implements Tool {
     133  static final String RANDOM_SEEK_SCAN = 
    "randomSeekScan";
     134  static final String RANDOM_READ = 
    "randomRead";
    -135  private static final Logger LOG = 
    LoggerFactory.getLogger(PerformanceEvaluation.class.getName());
    -136  private static final ObjectMapper 
    MAPPER = new ObjectMapper();
    -137  static {
    -138
    MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true);
    -139  }
    -140
    -141  public static final String TABLE_NAME = 
    "TestTable";
    -142  public static final String 
    FAMILY_NAME_BASE = "info";
    -143  public static final byte[] FAMILY_ZERO 
    = Bytes.toBytes("info0");
    -144  public static final byte[] COLUMN_ZERO 
    = Bytes.toBytes("" + 0);
    -145  public static final int 
    DEFAULT_VALUE_LENGTH = 1000;
    -146  public static final int ROW_LENGTH = 
    26;
    -147
    -148  private static final int ONE_GB = 1024 
    * 1024 * 1000;
    -149  private static final int 
    DEFAULT_ROWS_PER_GB = ONE_GB / DEFAULT_VALUE_LENGTH;
    -150  // TODO : should we make this 
    configurable
    -151  private static final int TAG_LENGTH = 
    256;
    -152  private static final DecimalFormat FMT 
    = new DecimalFormat("0.##");
    -153  private static final MathContext CXT = 
    MathContext.DECIMAL64;
    -154  private static final BigDecimal 
    MS_PER_SEC = BigDecimal.valueOf(1000);
    -155  private static final BigDecimal 
    BYTES_PER_MB = BigDecimal.valueOf(1024 * 1024);
    -156  private static final TestOptions 
    DEFAULT_OPTS = new TestOptions();
    -157
    -158  private static MapString, 
    CmdDescriptor COMMANDS = new TreeMap();
    -159  private static final Path PERF_EVAL_DIR 
    = new Path("performance_evaluation");
    -160
    -161  static {
    -162
    addCommandDescriptor(AsyncRandomReadTest.class, "asyncRandomRead",
    -163"Run async random read test");
    -164
    addCommandDescriptor(AsyncRandomWriteTest.class, "asyncRandomWrite",
    -165"Run async random write test");
    -166
    addCommandDescriptor(AsyncSequentialReadTest.class, "asyncSequentialRead",
    -167"Run async sequential read 
    test");
    -168
    addCommandDescriptor(AsyncSequentialWriteTest.class, "asyncSequentialWrite",
    -169"Run async sequential write 
    test");
    -170
    addCommandDescriptor(AsyncScanTest.class, "asyncScan",
    -171"Run async scan test (read every 
    row)");
    -172
    addCommandDescriptor(RandomReadTest.class, RANDOM_READ,
    -173  "Run random read test");
    -174
    addCommandDescriptor(RandomSeekScanTest.class, RANDOM_SEEK_SCAN,
    -175  "Run random seek and scan 100 
    test");
    -176
    addCommandDescriptor(RandomScanWithRange10Test.class, "scanRange10",
    -177  "Run random seek scan with both 
    start and stop row (max 10 rows)");
    -178
    addCommandDescriptor(RandomScanWithRange100Test.class, "scanRange100",
    -179  "Run random seek scan with both 
    start and stop row (max 100 rows)");
    -180
    addCommandDescriptor(RandomScanWithRange1000Test.class, "scanRange1000",
    -181  "Run random seek scan with both 
    start and stop row (max 1000 rows)");
    -182
    addCommandDescriptor(RandomScanWithRange1Test.class, "scanRange1",
    -183  "Run random seek scan with both 
    start and stop row (max 1 rows)");
    -184
    addCommandDescriptor(RandomWriteTest.class, "randomWrite",
    -185  "Run random write test");
    -186
    addCommandDescriptor(SequentialReadTest.class, "sequentialRead",
    -187  "Run sequential read test");
    -188
    addCommandDescriptor(SequentialWriteTest.class, "sequentialWrite",
    -189  "Run sequential write test");
    -190addCommandDescriptor(ScanTest.class, 
    "scan",
    -191  "Run scan test (read every 
    row)");
    -192
    addCommandDescriptor(FilteredScanTest.class, "filterScan",
    -193  "Run scan test using a filter to 
    find a specific row based on it's value " +
    -194  "(make sure to use --rows=20)");
    -195
    addCommandDescriptor(IncrementTest.class, "increment",
    -196  "Increment on each row; clients 
    overlap on keyspace so some concurrent operations");
    -197
    addCommandDescriptor(AppendTest.class, "append",
    -198  "Append on each row; clients 
    overlap on keyspace so some concurrent operations");
    -199
    addCommandDescriptor(CheckAndMutateTest.class, "checkAndMutate",
    -200  "CheckAndMutate on each row; 
    clients overlap on keyspace so some concurrent operations");
    -201
    addCommandDescriptor(CheckAndPutTest.class, "checkAndPut",
    

    [07/51] [partial] hbase-site git commit: Published site at 021f66d11d2cbb7308308093e29e69d6e7661ee9.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html 
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html
    index cdc0ca3..f434f88 100644
    --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html
    +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html
    @@ -27,109 +27,99 @@
     019
     020import 
    org.apache.hadoop.hbase.HBaseInterfaceAudience;
     021import 
    org.apache.yetus.audience.InterfaceAudience;
    -022
    -023/**
    -024 * Reports the data size part and total 
    heap space occupied by the MemStore.
    -025 * Read-only.
    +022/**
    +023 * Data structure of three longs.
    +024 * Convenient package in which to carry 
    current state of three counters.
    +025 * pImmutable!/p
     026 * @see MemStoreSizing
     027 */
     028@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
     029public class MemStoreSize {
    -030  // MemStore size tracks 3 sizes:
    -031  // (1) data size: the aggregated size 
    of all key-value not including meta data such as
    -032  // index, time range etc.
    -033  // (2) heap size: the aggregated size 
    of all data that is allocated on-heap including all
    -034  // key-values that reside on-heap and 
    the metadata that resides on-heap
    -035  // (3) off-heap size: the aggregated 
    size of all data that is allocated off-heap including all
    -036  // key-values that reside off-heap and 
    the metadata that resides off-heap
    -037  //
    -038  // 3 examples to illustrate their 
    usage:
    -039  // Consider a store with 100MB of 
    key-values allocated on-heap and 20MB of metadata allocated
    -040  // on-heap. The counters are 100MB, 
    120MB, 0, respectively.
    -041  // Consider a store with 100MB of 
    key-values allocated off-heap and 20MB of metadata
    -042  // allocated on-heap (e.g, CAM index). 
    The counters are 100MB, 20MB, 100MB, respectively.
    -043  // Consider a store with 100MB of 
    key-values from which 95MB are allocated off-heap and 5MB
    -044  // are allocated on-heap (e.g., due to 
    upserts) and 20MB of metadata from which 15MB allocated
    -045  // off-heap (e.g, CCM index) and 5MB 
    allocated on-heap (e.g, CSLM index in active).
    -046  // The counters are 100MB, 10MB, 
    110MB, respectively.
    -047
    -048  /**
    -049   *'dataSize' tracks the Cell's data 
    bytes size alone (Key bytes, value bytes). A cell's data can
    -050   * be in on heap or off heap area 
    depending on the MSLAB and its configuration to be using on heap
    -051   * or off heap LABs
    -052   */
    -053  protected volatile long dataSize;
    -054
    -055  /** 'heapSize' tracks all Cell's heap 
    size occupancy. This will include Cell POJO heap overhead.
    -056   * When Cells in on heap area, this 
    will include the cells data size as well.
    -057   */
    -058  protected volatile long heapSize;
    -059
    -060  /** off-heap size: the aggregated size 
    of all data that is allocated off-heap including all
    -061   * key-values that reside off-heap and 
    the metadata that resides off-heap
    -062   */
    -063  protected volatile long offHeapSize;
    -064
    -065  public MemStoreSize() {
    -066this(0L, 0L, 0L);
    -067  }
    -068
    -069  public MemStoreSize(long dataSize, long 
    heapSize, long offHeapSize) {
    -070this.dataSize = dataSize;
    -071this.heapSize = heapSize;
    -072this.offHeapSize = offHeapSize;
    -073  }
    -074
    -075  protected MemStoreSize(MemStoreSize 
    memStoreSize) {
    -076this.dataSize = 
    memStoreSize.dataSize;
    -077this.heapSize = 
    memStoreSize.heapSize;
    -078this.offHeapSize = 
    memStoreSize.offHeapSize;
    -079  }
    -080  public boolean isEmpty() {
    -081return this.dataSize == 0  
    this.heapSize == 0  this.offHeapSize == 0;
    +030  /**
    +031   *'dataSize' tracks the Cell's data 
    bytes size alone (Key bytes, value bytes). A cell's data can
    +032   * be in on heap or off heap area 
    depending on the MSLAB and its configuration to be using on
    +033   * heap or off heap LABs
    +034   */
    +035  private final long dataSize;
    +036
    +037  /**'getHeapSize' tracks all Cell's heap 
    size occupancy. This will include Cell POJO heap overhead.
    +038   * When Cells in on heap area, this 
    will include the cells data size as well.
    +039   */
    +040  private final long heapSize;
    +041
    +042  /** off-heap size: the aggregated size 
    of all data that is allocated off-heap including all
    +043   * key-values that reside off-heap and 
    the metadata that resides off-heap
    +044   */
    +045  private final long offHeapSize;
    +046
    +047  /**
    +048   * Package private constructor.
    +049   */
    +050  MemStoreSize() {
    +051this(0L, 0L, 0L);
    +052  }
    +053
    +054  /**
    +055   * Package private constructor.
    +056   */
    +057  MemStoreSize(long dataSize, long 
    heapSize, long offHeapSize) {
    +058this.dataSize = dataSize;
    +059this.heapSize = heapSize;
    +060this.offHeapSize = offHeapSize;
    +061  }
    +062
    +063  /**
    +064   * Package private constructor.
    +065   */
    +066  

    [07/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
    --
    diff --git 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
     
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
    index 2510283..418c60c 100644
    --- 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
    +++ 
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
    @@ -77,77 +77,77 @@
     069import 
    org.apache.hadoop.hbase.client.RowMutations;
     070import 
    org.apache.hadoop.hbase.client.Scan;
     071import 
    org.apache.hadoop.hbase.client.Table;
    -072import 
    org.apache.hadoop.hbase.filter.BinaryComparator;
    -073import 
    org.apache.hadoop.hbase.filter.Filter;
    -074import 
    org.apache.hadoop.hbase.filter.FilterAllFilter;
    -075import 
    org.apache.hadoop.hbase.filter.FilterList;
    -076import 
    org.apache.hadoop.hbase.filter.PageFilter;
    -077import 
    org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
    -078import 
    org.apache.hadoop.hbase.filter.WhileMatchFilter;
    -079import 
    org.apache.hadoop.hbase.io.compress.Compression;
    -080import 
    org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
    -081import 
    org.apache.hadoop.hbase.io.hfile.RandomDistribution;
    -082import 
    org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
    -083import 
    org.apache.hadoop.hbase.regionserver.BloomType;
    -084import 
    org.apache.hadoop.hbase.regionserver.CompactingMemStore;
    -085import 
    org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration;
    -086import 
    org.apache.hadoop.hbase.trace.SpanReceiverHost;
    -087import 
    org.apache.hadoop.hbase.trace.TraceUtil;
    -088import 
    org.apache.hadoop.hbase.util.ByteArrayHashKey;
    -089import 
    org.apache.hadoop.hbase.util.Bytes;
    -090import 
    org.apache.hadoop.hbase.util.Hash;
    -091import 
    org.apache.hadoop.hbase.util.MurmurHash;
    -092import 
    org.apache.hadoop.hbase.util.Pair;
    -093import 
    org.apache.hadoop.hbase.util.YammerHistogramUtils;
    -094import 
    org.apache.hadoop.io.LongWritable;
    -095import org.apache.hadoop.io.Text;
    -096import org.apache.hadoop.mapreduce.Job;
    -097import 
    org.apache.hadoop.mapreduce.Mapper;
    -098import 
    org.apache.hadoop.mapreduce.lib.input.NLineInputFormat;
    -099import 
    org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
    -100import 
    org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer;
    -101import org.apache.hadoop.util.Tool;
    -102import 
    org.apache.hadoop.util.ToolRunner;
    -103import 
    org.apache.htrace.core.ProbabilitySampler;
    -104import org.apache.htrace.core.Sampler;
    -105import 
    org.apache.htrace.core.TraceScope;
    -106import 
    org.apache.yetus.audience.InterfaceAudience;
    -107import org.slf4j.Logger;
    -108import org.slf4j.LoggerFactory;
    -109import 
    org.apache.hbase.thirdparty.com.google.common.base.MoreObjects;
    -110import 
    org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
    -111
    -112/**
    -113 * Script used evaluating HBase 
    performance and scalability.  Runs a HBase
    -114 * client that steps through one of a set 
    of hardcoded tests or 'experiments'
    -115 * (e.g. a random reads test, a random 
    writes test, etc.). Pass on the
    -116 * command-line which test to run and how 
    many clients are participating in
    -117 * this experiment. Run {@code 
    PerformanceEvaluation --help} to obtain usage.
    -118 *
    -119 * pThis class sets up and runs 
    the evaluation programs described in
    -120 * Section 7, iPerformance 
    Evaluation/i, of the a
    -121 * 
    href="http://labs.google.com/papers/bigtable.html"Bigtable/a;
    -122 * paper, pages 8-10.
    -123 *
    -124 * pBy default, runs as a 
    mapreduce job where each mapper runs a single test
    -125 * client. Can also run as a 
    non-mapreduce, multithreaded application by
    -126 * specifying {@code --nomapred}. Each 
    client does about 1GB of data, unless
    -127 * specified otherwise.
    -128 */
    -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
    -130public class PerformanceEvaluation 
    extends Configured implements Tool {
    -131  static final String RANDOM_SEEK_SCAN = 
    "randomSeekScan";
    -132  static final String RANDOM_READ = 
    "randomRead";
    -133  private static final Logger LOG = 
    LoggerFactory.getLogger(PerformanceEvaluation.class.getName());
    -134  private static final ObjectMapper 
    MAPPER = new ObjectMapper();
    -135  static {
    -136
    MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true);
    -137  }
    -138
    -139  public static final String TABLE_NAME = 
    "TestTable";
    -140  public static final byte[] FAMILY_NAME 
    = Bytes.toBytes("info");
    -141  public static final byte [] COLUMN_ZERO 
    = Bytes.toBytes("" + 0);
    -142  public static final byte [] 
    QUALIFIER_NAME = COLUMN_ZERO;
    +072import 
    org.apache.hadoop.hbase.client.metrics.ScanMetrics;
    +073import 
    org.apache.hadoop.hbase.filter.BinaryComparator;
    +074import 
    org.apache.hadoop.hbase.filter.Filter;
    +075import 
    

    [07/51] [partial] hbase-site git commit: Published site at 87f5b5f3411d96c31b4cb61b9a57ced22be91d1f.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
    --
    diff --git 
    a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
     
    b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
    index 7303f51..c604218 100644
    --- 
    a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
    +++ 
    b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
    @@ -32,9 +32,10 @@
     TestReplicationThrottler
     TestReplicator
     TestReplicator.FailureInjectingReplicationEndpointForTest
    -TestReplicator.FailureInjectingReplicationEndpointForTest.FailureInjectingBlockingInterface
     TestReplicator.ReplicationEndpointForTest
     TestSerialReplicationChecker
    +TestSerialReplicationEndpoint
    +TestSerialReplicationEndpoint.TestEndpoint
     TestSourceFSConfigurationProvider
     TestWALEntrySinkFilter
     TestWALEntrySinkFilter.DevNullConnection
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
    --
    diff --git 
    a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
     
    b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
    index 9b27b42..fd9dcc8 100644
    --- 
    a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
    +++ 
    b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
    @@ -173,40 +173,44 @@
     
     
     
    -TestReplicator.FailureInjectingReplicationEndpointForTest.FailureInjectingBlockingInterface
    +TestReplicator.ReplicationEndpointForTest
     
     
    -
    -TestReplicator.ReplicationEndpointForTest
    +
    +TestSerialReplicationChecker
     
     
     
    -TestSerialReplicationChecker
    +TestSerialReplicationEndpoint
     
     
     
    -TestSourceFSConfigurationProvider
    +TestSerialReplicationEndpoint.TestEndpoint
     
     
     
    +TestSourceFSConfigurationProvider
    +
    +
    +
     TestWALEntrySinkFilter
     
     Simple test of sink-side wal entry filter facility.
     
     
    -
    +
     TestWALEntrySinkFilter.DevNullConnection
     
     A DevNull Connection whose only purpose is checking what 
    edits made it through.
     
     
    -
    +
     TestWALEntrySinkFilter.IfTimeIsGreaterThanBOUNDARYWALEntrySinkFilterImpl
     
     Simple filter that will filter out any entry wholse 
    writeTime is <= 5.
     
     
    -
    +
     TestWALEntryStream
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
    --
    diff --git 
    a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
     
    b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
    index f7c6455..b47d481 100644
    --- 
    a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
    +++ 
    b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
    @@ -94,6 +94,7 @@
     org.apache.hadoop.hbase.replication.regionserver.TestReplicator.FailureInjectingReplicationEndpointForTest
     
     
    +org.apache.hadoop.hbase.replication.regionserver.TestSerialReplicationEndpoint.TestEndpoint
     
     
     
    @@ -138,13 +139,8 @@
     
     org.apache.hadoop.hbase.replication.regionserver.TestReplicationSourceManager.DummyServer 
    (implements org.apache.hadoop.hbase.Server)
     org.apache.hadoop.hbase.replication.regionserver.TestReplicationThrottler
    -org.apache.hadoop.hbase.replication.regionserver.TestReplicator.FailureInjectingReplicationEndpointForTest.FailureInjectingBlockingInterface
     (implements 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface)
    -org.apache.hadoop.hbase.replication.regionserver.TestReplicator.ReplicationEndpointForTest.ReplicatorForTest
    -
    -org.apache.hadoop.hbase.replication.regionserver.TestReplicator.FailureInjectingReplicationEndpointForTest.FailureInjectingReplicatorForTest
    -
    -
     org.apache.hadoop.hbase.replication.regionserver.TestSerialReplicationChecker
    +org.apache.hadoop.hbase.replication.regionserver.TestSerialReplicationEndpoint
     org.apache.hadoop.hbase.replication.regionserver.TestSourceFSConfigurationProvider (implements 
    org.apache.hadoop.hbase.replication.regionserver.SourceFSConfigurationProvider)
     org.apache.hadoop.hbase.replication.regionserver.TestWALEntrySinkFilter
     org.apache.hadoop.hbase.replication.regionserver.TestWALEntrySinkFilter.DevNullConnection 
    (implements org.apache.hadoop.hbase.client.Connection)
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-use.html
    --
    diff --git 
    a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-use.html
     
    

    [07/51] [partial] hbase-site git commit: Published site at 2912c953551bedbfbf30c32c156ed7bb187d54c3.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
    index 8302e28..c370eb9 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
    @@ -2113,3031 +2113,3033 @@
     2105
    errors.reportError(ERROR_CODE.ORPHAN_TABLE_STATE,
     2106tableName + " unable to 
    delete dangling table state " + tableState);
     2107  }
    -2108} else {
    -2109  
    errors.reportError(ERROR_CODE.ORPHAN_TABLE_STATE,
    -2110  tableName + " has dangling 
    table state " + tableState);
    -2111}
    -2112  }
    -2113}
    -2114// check that all tables have 
    states
    -2115for (TableName tableName : 
    tablesInfo.keySet()) {
    -2116  if (isTableIncluded(tableName) 
     !tableStates.containsKey(tableName)) {
    -2117if (fixMeta) {
    -2118  
    MetaTableAccessor.updateTableState(connection, tableName, 
    TableState.State.ENABLED);
    -2119  TableState newState = 
    MetaTableAccessor.getTableState(connection, tableName);
    -2120  if (newState == null) {
    -2121
    errors.reportError(ERROR_CODE.NO_TABLE_STATE,
    -2122"Unable to change state 
    for table " + tableName + " in meta ");
    -2123  }
    -2124} else {
    -2125  
    errors.reportError(ERROR_CODE.NO_TABLE_STATE,
    -2126  tableName + " has no state 
    in meta ");
    -2127}
    -2128  }
    -2129}
    -2130  }
    -2131
    -2132  private void preCheckPermission() 
    throws IOException, AccessDeniedException {
    -2133if 
    (shouldIgnorePreCheckPermission()) {
    -2134  return;
    -2135}
    -2136
    -2137Path hbaseDir = 
    FSUtils.getRootDir(getConf());
    -2138FileSystem fs = 
    hbaseDir.getFileSystem(getConf());
    -2139UserProvider userProvider = 
    UserProvider.instantiate(getConf());
    -2140UserGroupInformation ugi = 
    userProvider.getCurrent().getUGI();
    -2141FileStatus[] files = 
    fs.listStatus(hbaseDir);
    -2142for (FileStatus file : files) {
    -2143  try {
    -2144FSUtils.checkAccess(ugi, file, 
    FsAction.WRITE);
    -2145  } catch (AccessDeniedException 
    ace) {
    -2146LOG.warn("Got 
    AccessDeniedException when preCheckPermission ", ace);
    -2147
    errors.reportError(ERROR_CODE.WRONG_USAGE, "Current user " + 
    ugi.getUserName()
    -2148  + " does not have write perms 
    to " + file.getPath()
    -2149  + ". Please rerun hbck as hdfs 
    user " + file.getOwner());
    -2150throw ace;
    -2151  }
    -2152}
    -2153  }
    -2154
    -2155  /**
    -2156   * Deletes region from meta table
    -2157   */
    -2158  private void deleteMetaRegion(HbckInfo 
    hi) throws IOException {
    -2159
    deleteMetaRegion(hi.metaEntry.getRegionName());
    -2160  }
    -2161
    -2162  /**
    -2163   * Deletes region from meta table
    -2164   */
    -2165  private void deleteMetaRegion(byte[] 
    metaKey) throws IOException {
    -2166Delete d = new Delete(metaKey);
    -2167meta.delete(d);
    -2168LOG.info("Deleted " + 
    Bytes.toString(metaKey) + " from META" );
    -2169  }
    -2170
    -2171  /**
    -2172   * Reset the split parent region info 
    in meta table
    -2173   */
    -2174  private void resetSplitParent(HbckInfo 
    hi) throws IOException {
    -2175RowMutations mutations = new 
    RowMutations(hi.metaEntry.getRegionName());
    -2176Delete d = new 
    Delete(hi.metaEntry.getRegionName());
    -2177
    d.addColumn(HConstants.CATALOG_FAMILY, HConstants.SPLITA_QUALIFIER);
    -2178
    d.addColumn(HConstants.CATALOG_FAMILY, HConstants.SPLITB_QUALIFIER);
    -2179mutations.add(d);
    -2180
    -2181RegionInfo hri = 
    RegionInfoBuilder.newBuilder(hi.metaEntry)
    -2182.setOffline(false)
    -2183.setSplit(false)
    -2184.build();
    -2185Put p = 
    MetaTableAccessor.makePutFromRegionInfo(hri, 
    EnvironmentEdgeManager.currentTime());
    -2186mutations.add(p);
    -2187
    -2188meta.mutateRow(mutations);
    -2189LOG.info("Reset split parent " + 
    hi.metaEntry.getRegionNameAsString() + " in META" );
    -2190  }
    -2191
    -2192  /**
    -2193   * This backwards-compatibility 
    wrapper for permanently offlining a region
    -2194   * that should not be alive.  If the 
    region server does not support the
    -2195   * "offline" method, it will use the 
    closest unassign method instead.  This
    -2196   * will basically work until one 
    attempts to disable or delete the affected
    -2197   * table.  The problem has to do with 
    in-memory only master state, so
    -2198   * restarting the HMaster or failing 
    over to another should fix this.
    -2199   */
    -2200  private void offline(byte[] 
    regionName) throws IOException {
    -2201String regionString = 
    

    [07/51] [partial] hbase-site git commit: Published site at 2a2258656b2fcd92b967131b6c1f037363553bc4.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
    index 79bf967..c8b113b 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
    @@ -115,3514 +115,3517 @@
     107import 
    org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer;
     108import 
    org.apache.hadoop.hbase.master.balancer.ClusterStatusChore;
     109import 
    org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory;
    -110import 
    org.apache.hadoop.hbase.master.cleaner.HFileCleaner;
    -111import 
    org.apache.hadoop.hbase.master.cleaner.LogCleaner;
    -112import 
    org.apache.hadoop.hbase.master.cleaner.ReplicationBarrierCleaner;
    -113import 
    org.apache.hadoop.hbase.master.locking.LockManager;
    -114import 
    org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
    -115import 
    org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType;
    -116import 
    org.apache.hadoop.hbase.master.normalizer.RegionNormalizer;
    -117import 
    org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore;
    -118import 
    org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory;
    -119import 
    org.apache.hadoop.hbase.master.procedure.CreateTableProcedure;
    -120import 
    org.apache.hadoop.hbase.master.procedure.DeleteNamespaceProcedure;
    -121import 
    org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure;
    -122import 
    org.apache.hadoop.hbase.master.procedure.DisableTableProcedure;
    -123import 
    org.apache.hadoop.hbase.master.procedure.EnableTableProcedure;
    -124import 
    org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants;
    -125import 
    org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
    -126import 
    org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler;
    -127import 
    org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
    -128import 
    org.apache.hadoop.hbase.master.procedure.ModifyTableProcedure;
    -129import 
    org.apache.hadoop.hbase.master.procedure.ProcedurePrepareLatch;
    -130import 
    org.apache.hadoop.hbase.master.procedure.RecoverMetaProcedure;
    -131import 
    org.apache.hadoop.hbase.master.procedure.TruncateTableProcedure;
    -132import 
    org.apache.hadoop.hbase.master.replication.AddPeerProcedure;
    -133import 
    org.apache.hadoop.hbase.master.replication.DisablePeerProcedure;
    -134import 
    org.apache.hadoop.hbase.master.replication.EnablePeerProcedure;
    -135import 
    org.apache.hadoop.hbase.master.replication.ModifyPeerProcedure;
    -136import 
    org.apache.hadoop.hbase.master.replication.RemovePeerProcedure;
    -137import 
    org.apache.hadoop.hbase.master.replication.ReplicationPeerManager;
    -138import 
    org.apache.hadoop.hbase.master.replication.UpdatePeerConfigProcedure;
    -139import 
    org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
    -140import 
    org.apache.hadoop.hbase.mob.MobConstants;
    -141import 
    org.apache.hadoop.hbase.monitoring.MemoryBoundedLogMessageBuffer;
    -142import 
    org.apache.hadoop.hbase.monitoring.MonitoredTask;
    -143import 
    org.apache.hadoop.hbase.monitoring.TaskMonitor;
    -144import 
    org.apache.hadoop.hbase.procedure.MasterProcedureManagerHost;
    -145import 
    org.apache.hadoop.hbase.procedure.flush.MasterFlushTableProcedureManager;
    -146import 
    org.apache.hadoop.hbase.procedure2.LockedResource;
    -147import 
    org.apache.hadoop.hbase.procedure2.Procedure;
    -148import 
    org.apache.hadoop.hbase.procedure2.ProcedureEvent;
    -149import 
    org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
    -150import 
    org.apache.hadoop.hbase.procedure2.RemoteProcedureDispatcher.RemoteProcedure;
    -151import 
    org.apache.hadoop.hbase.procedure2.RemoteProcedureException;
    -152import 
    org.apache.hadoop.hbase.procedure2.store.wal.WALProcedureStore;
    -153import 
    org.apache.hadoop.hbase.quotas.MasterQuotaManager;
    -154import 
    org.apache.hadoop.hbase.quotas.MasterSpaceQuotaObserver;
    -155import 
    org.apache.hadoop.hbase.quotas.QuotaObserverChore;
    -156import 
    org.apache.hadoop.hbase.quotas.QuotaUtil;
    -157import 
    org.apache.hadoop.hbase.quotas.SnapshotQuotaObserverChore;
    -158import 
    org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshotNotifier;
    -159import 
    org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshotNotifierFactory;
    -160import 
    org.apache.hadoop.hbase.regionserver.DefaultStoreEngine;
    -161import 
    org.apache.hadoop.hbase.regionserver.HRegionServer;
    -162import 
    org.apache.hadoop.hbase.regionserver.HStore;
    -163import 
    org.apache.hadoop.hbase.regionserver.RSRpcServices;
    -164import 
    org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
    -165import 
    org.apache.hadoop.hbase.regionserver.RegionSplitPolicy;
    -166import 
    org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy;
    -167import 
    

    [07/51] [partial] hbase-site git commit: Published site at e468b4022f76688851b3e0c34722f01a56bd624f.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/devapidocs/org/apache/hadoop/hbase/client/HTable.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/client/HTable.html 
    b/devapidocs/org/apache/hadoop/hbase/client/HTable.html
    index 44fcf94..dd6854b 100644
    --- a/devapidocs/org/apache/hadoop/hbase/client/HTable.html
    +++ b/devapidocs/org/apache/hadoop/hbase/client/HTable.html
    @@ -115,7 +115,7 @@ var activeTableTab = "activeTableTab";
     
     @InterfaceAudience.Private
      @InterfaceStability.Stable
    -public class HTable
    +public class HTable
     extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     implements Table
     An implementation of Table. 
    Used to communicate with a single HBase table.
    @@ -515,29 +515,32 @@ implements 
     
     private boolean
    -doCheckAndDelete(byte[]row,
    +doCheckAndDelete(byte[]row,
     byte[]family,
     byte[]qualifier,
     https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">StringopName,
     byte[]value,
    +TimeRangetimeRange,
     Deletedelete)
     
     
     private boolean
    -doCheckAndMutate(byte[]row,
    +doCheckAndMutate(byte[]row,
     byte[]family,
     byte[]qualifier,
     https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">StringopName,
     byte[]value,
    +TimeRangetimeRange,
     RowMutationsrm)
     
     
     private boolean
    -doCheckAndPut(byte[]row,
    +doCheckAndPut(byte[]row,
      byte[]family,
      byte[]qualifier,
      https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">StringopName,
      byte[]value,
    + TimeRangetimeRange,
      Putput)
     
     
    @@ -832,7 +835,7 @@ implements 
     
     LOG
    -private static finalorg.slf4j.Logger LOG
    +private static finalorg.slf4j.Logger LOG
     
     
     
    @@ -841,7 +844,7 @@ implements 
     
     DEFAULT_CONSISTENCY
    -private static finalConsistency DEFAULT_CONSISTENCY
    +private static finalConsistency DEFAULT_CONSISTENCY
     
     
     
    @@ -850,7 +853,7 @@ implements 
     
     connection
    -private finalClusterConnection connection
    +private finalClusterConnection connection
     
     
     
    @@ -859,7 +862,7 @@ implements 
     
     tableName
    -private finalTableName tableName
    +private finalTableName tableName
     
     
     
    @@ -868,7 +871,7 @@ implements 
     
     configuration
    -private finalorg.apache.hadoop.conf.Configuration configuration
    +private finalorg.apache.hadoop.conf.Configuration configuration
     
     
     
    @@ -877,7 +880,7 @@ implements 
     
     connConfiguration
    -private finalConnectionConfiguration connConfiguration
    +private finalConnectionConfiguration connConfiguration
     
     
     
    @@ -886,7 +889,7 @@ implements 
     
     closed
    -privateboolean closed
    +privateboolean closed
     
     
     
    @@ -895,7 +898,7 @@ implements 
     
     scannerCaching
    -private finalint scannerCaching
    +private finalint scannerCaching
     
     
     
    @@ -904,7 +907,7 @@ implements 
     
     scannerMaxResultSize
    -private finallong scannerMaxResultSize
    +private finallong scannerMaxResultSize
     
     
     
    @@ -913,7 +916,7 @@ implements 
     
     pool
    -private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
     title="class or interface in java.util.concurrent">ExecutorService pool
    +private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
     title="class or interface in java.util.concurrent">ExecutorService pool
     
     
     
    @@ -922,7 +925,7 @@ implements 
     
     operationTimeoutMs
    -privateint operationTimeoutMs
    +privateint operationTimeoutMs
     
     
     
    @@ -931,7 +934,7 @@ implements 
     
     rpcTimeoutMs
    -private finalint rpcTimeoutMs
    +private finalint rpcTimeoutMs
     
     
     
    @@ -940,7 +943,7 @@ implements 
     
     readRpcTimeoutMs
    -privateint readRpcTimeoutMs
    +privateint readRpcTimeoutMs
     
     
     
    @@ -949,7 +952,7 @@ implements 
     
     writeRpcTimeoutMs
    -privateint writeRpcTimeoutMs
    +privateint writeRpcTimeoutMs
     
     
     
    @@ -958,7 +961,7 @@ implements 
     
     cleanupPoolOnClose
    -private finalboolean cleanupPoolOnClose
    +private finalboolean cleanupPoolOnClose
     
     
     
    @@ -967,7 +970,7 @@ implements 
     
     locator
    -private finalHRegionLocator locator
    +private finalHRegionLocator locator
     
     
     
    @@ -976,7 +979,7 @@ implements 
     
     multiAp
    -AsyncProcess multiAp
    +AsyncProcess multiAp
     The Async process for batch
     
     
    @@ -986,7 +989,7 @@ implements 
     
     rpcCallerFactory
    -private finalRpcRetryingCallerFactory 
    rpcCallerFactory
    +private finalRpcRetryingCallerFactory 
    rpcCallerFactory
     
     
     
    @@ -995,7 +998,7 @@ implements 
     
     rpcControllerFactory
    -private finalRpcControllerFactory rpcControllerFactory
    +private finalRpcControllerFactory rpcControllerFactory
     
     
     
    @@ -1013,7 +1016,7 @@ implements 
     HTable
     

    [07/51] [partial] hbase-site git commit: Published site at 64061f896fe21512504e3886a400759e88b519da.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html 
    b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
    index de426ec..c0b3f3f 100644
    --- a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
    +++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
    @@ -168,39 +168,27 @@
     
     
     void
    -CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    -  Cacheablebuf)
    -
    -
    -void
     BlockCache.cacheBlock(BlockCacheKeycacheKey,
       Cacheablebuf)
     Add block to cache (defaults to not in-memory).
     
     
    -
    +
     void
     LruBlockCache.cacheBlock(BlockCacheKeycacheKey,
       Cacheablebuf)
     Cache the block with the specified name and buffer.
     
     
    -
    -void
    -MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    -  Cacheablebuf)
    -
     
     void
    -CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    -  Cacheablebuf,
    -  booleaninMemory)
    +CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    +  Cacheablebuf)
     
     
     void
    -InclusiveCombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    -  Cacheablebuf,
    -  booleaninMemory)
    +MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    +  Cacheablebuf)
     
     
     void
    @@ -220,6 +208,18 @@
     
     
     void
    +CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    +  Cacheablebuf,
    +  booleaninMemory)
    +
    +
    +void
    +InclusiveCombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    +  Cacheablebuf,
    +  booleaninMemory)
    +
    +
    +void
     MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
       Cacheablebuf,
       booleaninMemory)
    @@ -232,21 +232,21 @@
     
     
     boolean
    -CombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
    +BlockCache.evictBlock(BlockCacheKeycacheKey)
    +Evict block from cache.
    +
     
     
     boolean
    -InclusiveCombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
    +LruBlockCache.evictBlock(BlockCacheKeycacheKey)
     
     
     boolean
    -BlockCache.evictBlock(BlockCacheKeycacheKey)
    -Evict block from cache.
    -
    +CombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
     
     
     boolean
    -LruBlockCache.evictBlock(BlockCacheKeycacheKey)
    +InclusiveCombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
     
     
     boolean
    @@ -254,35 +254,35 @@
     
     
     Cacheable
    -CombinedBlockCache.getBlock(BlockCacheKeycacheKey,
    +BlockCache.getBlock(BlockCacheKeycacheKey,
     booleancaching,
     booleanrepeat,
    -booleanupdateCacheMetrics)
    +booleanupdateCacheMetrics)
    +Fetch block from cache.
    +
     
     
     Cacheable
    -InclusiveCombinedBlockCache.getBlock(BlockCacheKeycacheKey,
    +LruBlockCache.getBlock(BlockCacheKeycacheKey,
     booleancaching,
     booleanrepeat,
    -booleanupdateCacheMetrics)
    +booleanupdateCacheMetrics)
    +Get the buffer of the block with the specified name.
    +
     
     
     Cacheable
    -BlockCache.getBlock(BlockCacheKeycacheKey,
    +CombinedBlockCache.getBlock(BlockCacheKeycacheKey,
     booleancaching,
     booleanrepeat,
    -booleanupdateCacheMetrics)
    -Fetch block from cache.
    -
    +booleanupdateCacheMetrics)
     
     
     Cacheable
    -LruBlockCache.getBlock(BlockCacheKeycacheKey,
    +InclusiveCombinedBlockCache.getBlock(BlockCacheKeycacheKey,
     booleancaching,
     booleanrepeat,
    -booleanupdateCacheMetrics)
    -Get the buffer of the block with the specified name.
    -
    +booleanupdateCacheMetrics)
     
     
     Cacheable
    @@ -308,11 +308,6 @@
     CombinedBlockCache.getRefCount(BlockCacheKeycacheKey)
     
     
    -void
    -CombinedBlockCache.returnBlock(BlockCacheKeycacheKey,
    -   Cacheableblock)
    -
    -
     default void
     BlockCache.returnBlock(BlockCacheKeycacheKey,
    Cacheableblock)
    @@ -320,6 +315,11 @@
      is over.
     
     
    +
    +void
    +CombinedBlockCache.returnBlock(BlockCacheKeycacheKey,
    +   Cacheableblock)
    +
     
     
     
    @@ -497,14 +497,14 @@
     
     
     void
    -BucketCache.BucketEntryGroup.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
     title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryblock)
    -
    -
    -void
     CachedEntryQueue.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
     title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryentry)
     Attempt to add the specified entry to this queue.
     
     
    +
    +void
    +BucketCache.BucketEntryGroup.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
     title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryblock)
    +
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockType.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockType.html 
    

    [07/51] [partial] hbase-site git commit: Published site at 4cb40e6d846ce1f28ffb40d388c9efb753197813.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html 
    b/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
    index 3b7627b..5f975db 100644
    --- a/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
    +++ b/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
    @@ -151,119 +151,119 @@
     
     
     Filter.ReturnCode
    -FilterListWithAND.filterCell(Cellc)
    +ColumnPrefixFilter.filterCell(Cellcell)
     
     
     Filter.ReturnCode
    -ValueFilter.filterCell(Cellc)
    +ColumnCountGetFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -SkipFilter.filterCell(Cellc)
    +RowFilter.filterCell(Cellv)
     
     
     Filter.ReturnCode
    -FamilyFilter.filterCell(Cellc)
    +FuzzyRowFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -ColumnPrefixFilter.filterCell(Cellcell)
    +Filter.filterCell(Cellc)
    +A way to filter based on the column family, column 
    qualifier and/or the column value.
    +
     
     
     Filter.ReturnCode
    -PageFilter.filterCell(Cellignored)
    +RandomRowFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -RowFilter.filterCell(Cellv)
    +FirstKeyOnlyFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -ColumnRangeFilter.filterCell(Cellc)
    +SkipFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -ColumnCountGetFilter.filterCell(Cellc)
    +TimestampsFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -MultipleColumnPrefixFilter.filterCell(Cellc)
    +ValueFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -ColumnPaginationFilter.filterCell(Cellc)
    +KeyOnlyFilter.filterCell(Cellignored)
     
     
     Filter.ReturnCode
    -DependentColumnFilter.filterCell(Cellc)
    +FamilyFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -FilterListWithOR.filterCell(Cellc)
    +QualifierFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -InclusiveStopFilter.filterCell(Cellc)
    +FilterList.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -KeyOnlyFilter.filterCell(Cellignored)
    +ColumnRangeFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -MultiRowRangeFilter.filterCell(Cellignored)
    +ColumnPaginationFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -Filter.filterCell(Cellc)
    -A way to filter based on the column family, column 
    qualifier and/or the column value.
    -
    +FilterListWithAND.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -FirstKeyOnlyFilter.filterCell(Cellc)
    +WhileMatchFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -WhileMatchFilter.filterCell(Cellc)
    +ColumnValueFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -FirstKeyValueMatchingQualifiersFilter.filterCell(Cellc)
    -Deprecated.
    -
    +MultiRowRangeFilter.filterCell(Cellignored)
     
     
     Filter.ReturnCode
    -TimestampsFilter.filterCell(Cellc)
    +PrefixFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -FuzzyRowFilter.filterCell(Cellc)
    +DependentColumnFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -FilterList.filterCell(Cellc)
    +FirstKeyValueMatchingQualifiersFilter.filterCell(Cellc)
    +Deprecated.
    +
     
     
     Filter.ReturnCode
    -RandomRowFilter.filterCell(Cellc)
    +PageFilter.filterCell(Cellignored)
     
     
     Filter.ReturnCode
    -PrefixFilter.filterCell(Cellc)
    +FilterListWithOR.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -SingleColumnValueFilter.filterCell(Cellc)
    +InclusiveStopFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -ColumnValueFilter.filterCell(Cellc)
    +MultipleColumnPrefixFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    -QualifierFilter.filterCell(Cellc)
    +SingleColumnValueFilter.filterCell(Cellc)
     
     
     Filter.ReturnCode
    @@ -279,158 +279,158 @@
     
     
     Filter.ReturnCode
    -ValueFilter.filterKeyValue(Cellc)
    +ColumnPrefixFilter.filterKeyValue(Cellc)
     Deprecated.
     
     
     
     Filter.ReturnCode
    -SkipFilter.filterKeyValue(Cellc)
    +ColumnCountGetFilter.filterKeyValue(Cellc)
     Deprecated.
     
     
     
     Filter.ReturnCode
    -FilterListBase.filterKeyValue(Cellc)
    +RowFilter.filterKeyValue(Cellc)
    +Deprecated.
    +
     
     
     Filter.ReturnCode
    -FamilyFilter.filterKeyValue(Cellc)
    +FuzzyRowFilter.filterKeyValue(Cellc)
     Deprecated.
     
     
     
     Filter.ReturnCode
    -ColumnPrefixFilter.filterKeyValue(Cellc)
    -Deprecated.
    +Filter.filterKeyValue(Cellc)
    +Deprecated.
    +As of release 2.0.0, this 
    will be removed in HBase 3.0.0.
    + Instead use filterCell(Cell)
    +
     
     
     
     Filter.ReturnCode
    -PageFilter.filterKeyValue(Cellc)
    +RandomRowFilter.filterKeyValue(Cellc)
     Deprecated.
     
     
     
     Filter.ReturnCode
    -RowFilter.filterKeyValue(Cellc)
    +FirstKeyOnlyFilter.filterKeyValue(Cellc)
     Deprecated.
     
     
     
     Filter.ReturnCode
    -ColumnRangeFilter.filterKeyValue(Cellc)
    +SkipFilter.filterKeyValue(Cellc)
     Deprecated.
     
     
     
     Filter.ReturnCode
    -ColumnCountGetFilter.filterKeyValue(Cellc)
    +TimestampsFilter.filterKeyValue(Cellc)
     Deprecated.
     
     
     
     Filter.ReturnCode
    -MultipleColumnPrefixFilter.filterKeyValue(Cellc)
    +ValueFilter.filterKeyValue(Cellc)
     Deprecated.
     
     
     
     Filter.ReturnCode
    -ColumnPaginationFilter.filterKeyValue(Cellc)
    +KeyOnlyFilter.filterKeyValue(Cellignored)
     Deprecated.
     
     
     
     Filter.ReturnCode
    

    [07/51] [partial] hbase-site git commit: Published site at 8ab7b20f48951d77945181024f5e15842bc253c4.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
    index ce22ca5..64099d9 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
    @@ -80,50 +80,57 @@
     072  }
     073
     074  @Override
    -075  public void sync() throws IOException 
    {
    +075  public void sync(boolean forceSync) 
    throws IOException {
     076FSDataOutputStream fsdos = 
    this.output;
     077if (fsdos == null) return; // Presume 
    closed
     078fsdos.flush();
    -079fsdos.hflush();
    -080  }
    -081
    -082  public FSDataOutputStream getStream() 
    {
    -083return this.output;
    +079if (forceSync) {
    +080  fsdos.hsync();
    +081} else {
    +082  fsdos.hflush();
    +083}
     084  }
     085
    -086  @SuppressWarnings("deprecation")
    -087  @Override
    -088  protected void initOutput(FileSystem 
    fs, Path path, boolean overwritable, int bufferSize,
    -089  short replication, long blockSize) 
    throws IOException, StreamLacksCapabilityException {
    -090this.output = 
    fs.createNonRecursive(path, overwritable, bufferSize, replication, blockSize,
    -091  null);
    -092// TODO Be sure to add a check for 
    hsync if this branch includes HBASE-19024
    -093if 
    (fs.getConf().getBoolean(CommonFSUtils.UNSAFE_STREAM_CAPABILITY_ENFORCE, true) 
    
    -094
    !(CommonFSUtils.hasCapability(output, "hflush"))) {
    -095  throw new 
    StreamLacksCapabilityException("hflush");
    -096}
    -097  }
    -098
    -099  @Override
    -100  protected long 
    writeMagicAndWALHeader(byte[] magic, WALHeader header) throws IOException {
    -101output.write(magic);
    -102header.writeDelimitedTo(output);
    -103return output.getPos();
    +086  public FSDataOutputStream getStream() 
    {
    +087return this.output;
    +088  }
    +089
    +090  @SuppressWarnings("deprecation")
    +091  @Override
    +092  protected void initOutput(FileSystem 
    fs, Path path, boolean overwritable, int bufferSize,
    +093  short replication, long blockSize) 
    throws IOException, StreamLacksCapabilityException {
    +094this.output = 
    fs.createNonRecursive(path, overwritable, bufferSize, replication, blockSize,
    +095  null);
    +096if 
    (fs.getConf().getBoolean(CommonFSUtils.UNSAFE_STREAM_CAPABILITY_ENFORCE, true)) 
    {
    +097  if 
    (!CommonFSUtils.hasCapability(output, "hflush")) {
    +098throw new 
    StreamLacksCapabilityException("hflush");
    +099  }
    +100  if 
    (!CommonFSUtils.hasCapability(output, "hsync")) {
    +101throw new 
    StreamLacksCapabilityException("hsync");
    +102  }
    +103}
     104  }
     105
     106  @Override
    -107  protected OutputStream 
    getOutputStreamForCellEncoder() {
    -108return this.output;
    -109  }
    -110
    -111  @Override
    -112  protected long 
    writeWALTrailerAndMagic(WALTrailer trailer, byte[] magic) throws IOException 
    {
    -113trailer.writeTo(output);
    -114
    output.writeInt(trailer.getSerializedSize());
    -115output.write(magic);
    -116return output.getPos();
    -117  }
    -118}
    +107  protected long 
    writeMagicAndWALHeader(byte[] magic, WALHeader header) throws IOException {
    +108output.write(magic);
    +109header.writeDelimitedTo(output);
    +110return output.getPos();
    +111  }
    +112
    +113  @Override
    +114  protected OutputStream 
    getOutputStreamForCellEncoder() {
    +115return this.output;
    +116  }
    +117
    +118  @Override
    +119  protected long 
    writeWALTrailerAndMagic(WALTrailer trailer, byte[] magic) throws IOException 
    {
    +120trailer.writeTo(output);
    +121
    output.writeInt(trailer.getSerializedSize());
    +122output.write(magic);
    +123return output.getPos();
    +124  }
    +125}
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html 
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html
    index 9457b4b..6a84f26 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html
    @@ -75,94 +75,105 @@
     067
     068  private Thread t;
     069
    -070  /**
    -071   * Call this method to clear old usage 
    and get it ready for new deploy.
    -072   * @param txid the new transaction id
    -073   * @param span current span, detached 
    from caller. Don't forget to attach it when resuming after a
    -074   *  call to {@link 
    #get(long)}.
    -075   * @return this
    -076   */
    -077  synchronized SyncFuture reset(long 
    txid) {
    -078if (t != null  t != 
    Thread.currentThread()) {
    -079  

    [07/51] [partial] hbase-site git commit: Published site at 00095a2ef9442e3fd86c04876c9d91f2f8b23ad8.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
    index 3bc66bb..97aa79c 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
    @@ -1435,459 +1435,460 @@
     1427   */
     1428  private void execProcedure(final 
    RootProcedureState procStack,
     1429  final 
    ProcedureTEnvironment procedure) {
    -1430
    Preconditions.checkArgument(procedure.getState() == ProcedureState.RUNNABLE);
    -1431
    -1432// Procedures can suspend 
    themselves. They skip out by throwing a ProcedureSuspendedException.
    -1433// The exception is caught below and 
    then we hurry to the exit without disturbing state. The
    -1434// idea is that the processing of 
    this procedure will be unsuspended later by an external event
    -1435// such the report of a region open. 
    TODO: Currently, its possible for two worker threads
    -1436// to be working on the same 
    procedure concurrently (locking in procedures is NOT about
    -1437// concurrency but about tying an 
    entity to a procedure; i.e. a region to a particular
    -1438// procedure instance). This can 
    make for issues if both threads are changing state.
    -1439// See 
    env.getProcedureScheduler().wakeEvent(regionNode.getProcedureEvent());
    -1440// in 
    RegionTransitionProcedure#reportTransition for example of Procedure putting
    -1441// itself back on the scheduler 
    making it possible for two threads running against
    -1442// the one Procedure. Might be ok if 
    they are both doing different, idempotent sections.
    -1443boolean suspended = false;
    -1444
    -1445// Whether to 're-' -execute; run 
    through the loop again.
    -1446boolean reExecute = false;
    -1447
    -1448ProcedureTEnvironment[] 
    subprocs = null;
    -1449do {
    -1450  reExecute = false;
    -1451  try {
    -1452subprocs = 
    procedure.doExecute(getEnvironment());
    -1453if (subprocs != null  
    subprocs.length == 0) {
    -1454  subprocs = null;
    -1455}
    -1456  } catch 
    (ProcedureSuspendedException e) {
    -1457if (LOG.isTraceEnabled()) {
    -1458  LOG.trace("Suspend " + 
    procedure);
    -1459}
    -1460suspended = true;
    -1461  } catch (ProcedureYieldException 
    e) {
    -1462if (LOG.isTraceEnabled()) {
    -1463  LOG.trace("Yield " + procedure 
    + ": " + e.getMessage(), e);
    -1464}
    -1465scheduler.yield(procedure);
    -1466return;
    -1467  } catch (InterruptedException e) 
    {
    -1468if (LOG.isTraceEnabled()) {
    -1469  LOG.trace("Yield interrupt " + 
    procedure + ": " + e.getMessage(), e);
    -1470}
    -1471
    handleInterruptedException(procedure, e);
    -1472scheduler.yield(procedure);
    -1473return;
    -1474  } catch (Throwable e) {
    -1475// Catch NullPointerExceptions 
    or similar errors...
    -1476String msg = "CODE-BUG: Uncaught 
    runtime exception: " + procedure;
    -1477LOG.error(msg, e);
    -1478procedure.setFailure(new 
    RemoteProcedureException(msg, e));
    -1479  }
    -1480
    -1481  if (!procedure.isFailed()) {
    -1482if (subprocs != null) {
    -1483  if (subprocs.length == 1 
     subprocs[0] == procedure) {
    -1484// Procedure returned 
    itself. Quick-shortcut for a state machine-like procedure;
    -1485// i.e. we go around this 
    loop again rather than go back out on the scheduler queue.
    -1486subprocs = null;
    -1487reExecute = true;
    -1488if (LOG.isTraceEnabled()) 
    {
    -1489  LOG.trace("Short-circuit 
    to next step on pid=" + procedure.getProcId());
    -1490}
    -1491  } else {
    -1492// Yield the current 
    procedure, and make the subprocedure runnable
    -1493// subprocs may come back 
    'null'.
    -1494subprocs = 
    initializeChildren(procStack, procedure, subprocs);
    -1495LOG.info("Initialized 
    subprocedures=" +
    -1496  (subprocs == null? null:
    -1497
    Stream.of(subprocs).map(e - "{" + e.toString() + "}").
    -1498
    collect(Collectors.toList()).toString()));
    -1499  }
    -1500} else if (procedure.getState() 
    == ProcedureState.WAITING_TIMEOUT) {
    -1501  if (LOG.isTraceEnabled()) {
    -1502LOG.trace("Added to 
    timeoutExecutor " + procedure);
    -1503  }
    -1504  
    timeoutExecutor.add(procedure);
    -1505} else if (!suspended) {
    -1506  // No subtask, so we are 
    done
    -1507  
    procedure.setState(ProcedureState.SUCCESS);
    

    [07/51] [partial] hbase-site git commit: Published site at 22f4def942f8a3367d0ca6598317e9b9a7d0cfcd.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
    --
    diff --git 
    a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html 
    b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
    index 3a91ea6..670bcf3 100644
    --- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
    +++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
    @@ -975,630 +975,634 @@
     
     
     static HBaseClassTestRule
    -TestImmutableHRegionInfo.CLASS_RULE
    +TestCoprocessorDescriptor.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestMutation.CLASS_RULE
    +TestImmutableHRegionInfo.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestRowComparator.CLASS_RULE
    +TestMutation.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestBufferedMutator.CLASS_RULE
    +TestRowComparator.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestClientScanner.CLASS_RULE
    +TestBufferedMutator.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestInterfaceAlign.CLASS_RULE
    +TestClientScanner.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestBufferedMutatorParams.CLASS_RULE
    +TestInterfaceAlign.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestSnapshotFromAdmin.CLASS_RULE
    +TestBufferedMutatorParams.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestIncrement.CLASS_RULE
    +TestSnapshotFromAdmin.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncProcessWithRegionException.CLASS_RULE
    +TestIncrement.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestHTableMultiplexerViaMocks.CLASS_RULE
    +TestAsyncProcessWithRegionException.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestColumnFamilyDescriptorBuilder.CLASS_RULE
    +TestHTableMultiplexerViaMocks.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestMetricsConnection.CLASS_RULE
    +TestColumnFamilyDescriptorBuilder.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestTableDescriptorBuilder.CLASS_RULE
    +TestMetricsConnection.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestImmutableHColumnDescriptor.CLASS_RULE
    +TestTableDescriptorBuilder.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestImmutableHTableDescriptor.CLASS_RULE
    +TestImmutableHColumnDescriptor.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestRetriesExhaustedWithDetailsException.CLASS_RULE
    +TestImmutableHTableDescriptor.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncProcess.CLASS_RULE
    +TestRetriesExhaustedWithDetailsException.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestProcedureFuture.CLASS_RULE
    +TestAsyncProcess.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestClientExponentialBackoff.CLASS_RULE
    +TestProcedureFuture.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestDeleteTimeStamp.CLASS_RULE
    +TestClientExponentialBackoff.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestOperation.CLASS_RULE
    +TestDeleteTimeStamp.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestMultiParallel.CLASS_RULE
    +TestOperation.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestSnapshotMetadata.CLASS_RULE
    +TestMultiParallel.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestFromClientSideNoCodec.CLASS_RULE
    +TestSnapshotMetadata.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncBufferMutator.CLASS_RULE
    +TestFromClientSideNoCodec.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestSizeFailures.CLASS_RULE
    +TestAsyncBufferMutator.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestGetProcedureResult.CLASS_RULE
    +TestSizeFailures.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncTableScanRenewLease.CLASS_RULE
    +TestGetProcedureResult.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestClientOperationInterrupt.CLASS_RULE
    +TestAsyncTableScanRenewLease.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestScanWithoutFetchingData.CLASS_RULE
    +TestClientOperationInterrupt.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestServerLoadDurability.CLASS_RULE
    +TestScanWithoutFetchingData.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncTableScan.CLASS_RULE
    +TestServerLoadDurability.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestMalformedCellFromClient.CLASS_RULE
    +TestAsyncTableScan.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncReplicationAdminApiWithClusters.CLASS_RULE
    +TestMalformedCellFromClient.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestCheckAndMutate.CLASS_RULE
    +TestAsyncReplicationAdminApiWithClusters.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestFastFail.CLASS_RULE
    +TestCheckAndMutate.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAppendFromClientSide.CLASS_RULE
    +TestFastFail.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestServerBusyException.CLASS_RULE
    +TestAppendFromClientSide.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestMobSnapshotFromClient.CLASS_RULE
    +TestServerBusyException.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestAsyncQuotaAdminApi.CLASS_RULE
    +TestMobSnapshotFromClient.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestSplitOrMergeStatus.CLASS_RULE
    +TestAsyncQuotaAdminApi.CLASS_RULE
     
     
     static HBaseClassTestRule
    -TestFromClientSideScanExcpetionWithCoprocessor.CLASS_RULE
    

    [07/51] [partial] hbase-site git commit: Published site at 31da4d0bce69b3a47066a5df675756087ce4dc60.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html 
    b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
    index ec5a869..84f7079 100644
    --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
    +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
    @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab";
     
     
     @InterfaceAudience.Private
    -public class RSRpcServices
    +public class RSRpcServices
     extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     implements HBaseRPCErrorHandler, 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface,
     
    org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface,
     PriorityFunction, ConfigurationObserver
     Implements the regionserver RPC services.
    @@ -191,7 +191,7 @@ implements Field and Description
     
     
    -private AccessChecker
    +protected AccessChecker
     accessChecker
     
     
    @@ -959,7 +959,7 @@ implements 
     
     LOG
    -protected static finalorg.slf4j.Logger LOG
    +protected static finalorg.slf4j.Logger LOG
     
     
     
    @@ -968,7 +968,7 @@ implements 
     
     REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS
    -public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS
    +public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS
     RPC scheduler to use for the region server.
     
     See Also:
    @@ -982,7 +982,7 @@ implements 
     
     REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
    -private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
    +private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
     Minimum allowable time limit delta (in milliseconds) that 
    can be enforced during scans. This
      configuration exists to prevent the scenario where a time limit is specified 
    to be so
      restrictive that the time limit is reached immediately (before any cells are 
    scanned).
    @@ -998,7 +998,7 @@ implements 
     
     DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
    -private static finallong DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
    +private static finallong DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
     Default value of REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
     
     See Also:
    @@ -1012,7 +1012,7 @@ implements 
     
     BATCH_ROWS_THRESHOLD_NAME
    -static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String BATCH_ROWS_THRESHOLD_NAME
    +static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String BATCH_ROWS_THRESHOLD_NAME
     Number of rows in a batch operation above which a warning 
    will be logged.
     
     See Also:
    @@ -1026,7 +1026,7 @@ implements 
     
     BATCH_ROWS_THRESHOLD_DEFAULT
    -static finalint BATCH_ROWS_THRESHOLD_DEFAULT
    +static finalint BATCH_ROWS_THRESHOLD_DEFAULT
     Default value of BATCH_ROWS_THRESHOLD_NAME
     
     See Also:
    @@ -1040,7 +1040,7 @@ implements 
     
     RESERVOIR_ENABLED_KEY
    -protected static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String RESERVOIR_ENABLED_KEY
    +protected static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String RESERVOIR_ENABLED_KEY
     
     See Also:
     Constant
     Field Values
    @@ -1053,7 +1053,7 @@ implements 
     
     requestCount
    -finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true;
     title="class or interface in java.util.concurrent.atomic">LongAdder requestCount
    +finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true;
     title="class or interface in java.util.concurrent.atomic">LongAdder requestCount
     
     
     
    @@ -1062,7 +1062,7 @@ implements 
     
     rpcGetRequestCount
    -finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true;
     title="class or interface in java.util.concurrent.atomic">LongAdder rpcGetRequestCount
    +finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true;
     title="class or interface in 

    [07/51] [partial] hbase-site git commit: Published site at 6b77786dfc46d25ac5bb5f1c8a4a9eb47b52a604.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html 
    b/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html
    index dbe6db5..0b18d24 100644
    --- a/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html
    +++ b/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html
    @@ -353,7 +353,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     
    -void
    +boolean
     expireServer(ServerNameserverName)
     
     
    @@ -1126,7 +1126,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     expireServer
    -publicvoidexpireServer(ServerNameserverName)
    +publicbooleanexpireServer(ServerNameserverName)
     
     
     
    @@ -1135,7 +1135,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     moveFromOnlineToDeadServers
    -publicvoidmoveFromOnlineToDeadServers(ServerNamesn)
    +publicvoidmoveFromOnlineToDeadServers(ServerNamesn)
     
     
     
    @@ -1144,7 +1144,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     processDeadServer
    -publicvoidprocessDeadServer(ServerNameserverName,
    +publicvoidprocessDeadServer(ServerNameserverName,
       booleanshouldSplitWal)
     
     
    @@ -1154,7 +1154,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     processQueuedDeadServers
    -voidprocessQueuedDeadServers()
    +voidprocessQueuedDeadServers()
     Process the servers which died during master's 
    initialization. It will be
      called after HMaster#assignMeta and AssignmentManager#joinCluster.
     
    @@ -1165,7 +1165,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     removeServerFromDrainList
    -publicbooleanremoveServerFromDrainList(ServerNamesn)
    +publicbooleanremoveServerFromDrainList(ServerNamesn)
     
     
     
    @@ -1174,7 +1174,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     addServerToDrainList
    -publicbooleanaddServerToDrainList(ServerNamesn)
    +publicbooleanaddServerToDrainList(ServerNamesn)
     Add the server to the drain list.
     
     Parameters:
    @@ -1190,7 +1190,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     newRpcController
    -privateHBaseRpcControllernewRpcController()
    +privateHBaseRpcControllernewRpcController()
     
     
     
    @@ -1199,7 +1199,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     sendRegionWarmup
    -publicvoidsendRegionWarmup(ServerNameserver,
    +publicvoidsendRegionWarmup(ServerNameserver,
      RegionInforegion)
     Sends a WARMUP RPC to the specified server to warmup the 
    specified region.
      
    @@ -1218,7 +1218,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     closeRegionSilentlyAndWait
    -public staticvoidcloseRegionSilentlyAndWait(ClusterConnectionconnection,
    +public staticvoidcloseRegionSilentlyAndWait(ClusterConnectionconnection,
       ServerNameserver,
       RegionInforegion,
       longtimeout)
    @@ -1239,7 +1239,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     getRsAdmin
    -publicorg.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterfacegetRsAdmin(ServerNamesn)
    +publicorg.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterfacegetRsAdmin(ServerNamesn)
    
     throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
     title="class or interface in java.io">IOException
     
     Parameters:
    @@ -1258,7 +1258,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     getMinToStart
    -privateintgetMinToStart()
    +privateintgetMinToStart()
     Calculate min necessary to start. This is not an absolute. 
    It is just
      a friction that will cause us hang around a bit longer waiting on
      RegionServers to check-in.
    @@ -1270,7 +1270,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     waitForRegionServers
    -publicvoidwaitForRegionServers(MonitoredTaskstatus)
    +publicvoidwaitForRegionServers(MonitoredTaskstatus)
       throws https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
     title="class or interface in java.lang">InterruptedException
     Wait for the region servers to report in.
      We will wait until one of this condition is met:
    @@ -1293,7 +1293,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
     
     
     getStrForMax
    -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">StringgetStrForMax(intmax)
    

    [07/51] [partial] hbase-site git commit: Published site at 1384da71375427b522b09f06862bb5d629cef52f.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html 
    b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
    index c0b3f3f..de426ec 100644
    --- a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
    +++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
    @@ -168,27 +168,39 @@
     
     
     void
    +CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    +  Cacheablebuf)
    +
    +
    +void
     BlockCache.cacheBlock(BlockCacheKeycacheKey,
       Cacheablebuf)
     Add block to cache (defaults to not in-memory).
     
     
    -
    +
     void
     LruBlockCache.cacheBlock(BlockCacheKeycacheKey,
       Cacheablebuf)
     Cache the block with the specified name and buffer.
     
     
    -
    +
     void
    -CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    +MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
       Cacheablebuf)
     
    +
    +void
    +CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    +  Cacheablebuf,
    +  booleaninMemory)
    +
     
     void
    -MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    -  Cacheablebuf)
    +InclusiveCombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    +  Cacheablebuf,
    +  booleaninMemory)
     
     
     void
    @@ -208,18 +220,6 @@
     
     
     void
    -CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    -  Cacheablebuf,
    -  booleaninMemory)
    -
    -
    -void
    -InclusiveCombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
    -  Cacheablebuf,
    -  booleaninMemory)
    -
    -
    -void
     MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
       Cacheablebuf,
       booleaninMemory)
    @@ -232,21 +232,21 @@
     
     
     boolean
    -BlockCache.evictBlock(BlockCacheKeycacheKey)
    -Evict block from cache.
    -
    +CombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
     
     
     boolean
    -LruBlockCache.evictBlock(BlockCacheKeycacheKey)
    +InclusiveCombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
     
     
     boolean
    -CombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
    +BlockCache.evictBlock(BlockCacheKeycacheKey)
    +Evict block from cache.
    +
     
     
     boolean
    -InclusiveCombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
    +LruBlockCache.evictBlock(BlockCacheKeycacheKey)
     
     
     boolean
    @@ -254,35 +254,35 @@
     
     
     Cacheable
    -BlockCache.getBlock(BlockCacheKeycacheKey,
    +CombinedBlockCache.getBlock(BlockCacheKeycacheKey,
     booleancaching,
     booleanrepeat,
    -booleanupdateCacheMetrics)
    -Fetch block from cache.
    -
    +booleanupdateCacheMetrics)
     
     
     Cacheable
    -LruBlockCache.getBlock(BlockCacheKeycacheKey,
    +InclusiveCombinedBlockCache.getBlock(BlockCacheKeycacheKey,
     booleancaching,
     booleanrepeat,
    -booleanupdateCacheMetrics)
    -Get the buffer of the block with the specified name.
    -
    +booleanupdateCacheMetrics)
     
     
     Cacheable
    -CombinedBlockCache.getBlock(BlockCacheKeycacheKey,
    +BlockCache.getBlock(BlockCacheKeycacheKey,
     booleancaching,
     booleanrepeat,
    -booleanupdateCacheMetrics)
    +booleanupdateCacheMetrics)
    +Fetch block from cache.
    +
     
     
     Cacheable
    -InclusiveCombinedBlockCache.getBlock(BlockCacheKeycacheKey,
    +LruBlockCache.getBlock(BlockCacheKeycacheKey,
     booleancaching,
     booleanrepeat,
    -booleanupdateCacheMetrics)
    +booleanupdateCacheMetrics)
    +Get the buffer of the block with the specified name.
    +
     
     
     Cacheable
    @@ -308,6 +308,11 @@
     CombinedBlockCache.getRefCount(BlockCacheKeycacheKey)
     
     
    +void
    +CombinedBlockCache.returnBlock(BlockCacheKeycacheKey,
    +   Cacheableblock)
    +
    +
     default void
     BlockCache.returnBlock(BlockCacheKeycacheKey,
    Cacheableblock)
    @@ -315,11 +320,6 @@
      is over.
     
     
    -
    -void
    -CombinedBlockCache.returnBlock(BlockCacheKeycacheKey,
    -   Cacheableblock)
    -
     
     
     
    @@ -497,13 +497,13 @@
     
     
     void
    -CachedEntryQueue.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
     title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryentry)
    -Attempt to add the specified entry to this queue.
    -
    +BucketCache.BucketEntryGroup.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
     title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryblock)
     
     
     void
    -BucketCache.BucketEntryGroup.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
     title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryblock)
    +CachedEntryQueue.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
     title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryentry)
    +Attempt to add the specified entry to this queue.
    +
     
     
     
    
    

    [07/51] [partial] hbase-site git commit: Published site at b7b86839250bf9b295ebc1948826f43a88736d6c.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html
    index c8098b1..019ceeb 100644
    --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html
    +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html
    @@ -216,245 +216,254 @@
     208  /**
     209   * @return HBase root log dir.
     210   */
    -211  public Path getWALRootDir() { return 
    this.walRootDir; }
    -212
    -213  /**
    -214   * @return HBase temp dir.
    -215   */
    -216  public Path getTempDir() {
    -217return this.tempdir;
    -218  }
    -219
    -220  /**
    -221   * @return The unique identifier 
    generated for this cluster
    -222   */
    -223  public ClusterId getClusterId() {
    -224return clusterId;
    -225  }
    -226
    -227  /**
    -228   * Get the rootdir.  Make sure its 
    wholesome and exists before returning.
    -229   * @param rd
    -230   * @param c
    -231   * @param fs
    -232   * @return hbase.rootdir (after checks 
    for existence and bootstrapping if
    -233   * needed populating the directory with 
    necessary bootup files).
    -234   * @throws IOException
    -235   */
    -236  private Path checkRootDir(final Path 
    rd, final Configuration c, final FileSystem fs)
    -237  throws IOException {
    -238// If FS is in safe mode wait till 
    out of it.
    -239FSUtils.waitOnSafeMode(c, 
    c.getInt(HConstants.THREAD_WAKE_FREQUENCY, 10 * 1000));
    -240
    -241// Filesystem is good. Go ahead and 
    check for hbase.rootdir.
    -242try {
    -243  if (!fs.exists(rd)) {
    -244fs.mkdirs(rd);
    -245// DFS leaves safe mode with 0 
    DNs when there are 0 blocks.
    -246// We used to handle this by 
    checking the current DN count and waiting until
    -247// it is nonzero. With security, 
    the check for datanode count doesn't work --
    -248// it is a privileged op. So 
    instead we adopt the strategy of the jobtracker
    -249// and simply retry file creation 
    during bootstrap indefinitely. As soon as
    -250// there is one datanode it will 
    succeed. Permission problems should have
    -251// already been caught by mkdirs 
    above.
    -252FSUtils.setVersion(fs, rd, 
    c.getInt(HConstants.THREAD_WAKE_FREQUENCY,
    -253  10 * 1000), 
    c.getInt(HConstants.VERSION_FILE_WRITE_ATTEMPTS,
    -254
    HConstants.DEFAULT_VERSION_FILE_WRITE_ATTEMPTS));
    -255  } else {
    -256if (!fs.isDirectory(rd)) {
    -257  throw new 
    IllegalArgumentException(rd.toString() + " is not a directory");
    -258}
    -259// as above
    -260FSUtils.checkVersion(fs, rd, 
    true, c.getInt(HConstants.THREAD_WAKE_FREQUENCY,
    -261  10 * 1000), 
    c.getInt(HConstants.VERSION_FILE_WRITE_ATTEMPTS,
    -262
    HConstants.DEFAULT_VERSION_FILE_WRITE_ATTEMPTS));
    -263  }
    -264} catch (DeserializationException de) 
    {
    -265  LOG.error(HBaseMarkers.FATAL, 
    "Please fix invalid configuration for "
    -266+ HConstants.HBASE_DIR, de);
    -267  IOException ioe = new 
    IOException();
    -268  ioe.initCause(de);
    -269  throw ioe;
    -270} catch (IllegalArgumentException 
    iae) {
    -271  LOG.error(HBaseMarkers.FATAL, 
    "Please fix invalid configuration for "
    -272+ HConstants.HBASE_DIR + " " + 
    rd.toString(), iae);
    -273  throw iae;
    -274}
    -275// Make sure cluster ID exists
    -276if (!FSUtils.checkClusterIdExists(fs, 
    rd, c.getInt(
    -277HConstants.THREAD_WAKE_FREQUENCY, 
    10 * 1000))) {
    -278  FSUtils.setClusterId(fs, rd, new 
    ClusterId(), c.getInt(HConstants.THREAD_WAKE_FREQUENCY, 10 * 1000));
    -279}
    -280clusterId = FSUtils.getClusterId(fs, 
    rd);
    -281
    -282// Make sure the meta region 
    directory exists!
    -283if (!FSUtils.metaRegionExists(fs, 
    rd)) {
    -284  bootstrap(rd, c);
    -285}
    -286
    -287// Create tableinfo-s for hbase:meta 
    if not already there.
    -288// assume, created table descriptor 
    is for enabling table
    -289// meta table is a system table, so 
    descriptors are predefined,
    -290// we should get them from 
    registry.
    -291FSTableDescriptors fsd = new 
    FSTableDescriptors(c, fs, rd);
    -292
    fsd.createTableDescriptor(fsd.get(TableName.META_TABLE_NAME));
    -293
    -294return rd;
    -295  }
    -296
    -297  /**
    -298   * Make sure the hbase temp directory 
    exists and is empty.
    -299   * NOTE that this method is only 
    executed once just after the master becomes the active one.
    -300   */
    -301  private void checkTempDir(final Path 
    tmpdir, final Configuration c, final FileSystem fs)
    -302  throws IOException {
    -303// If the temp directory exists, 
    clear the content (left over, from the previous run)
    -304if (fs.exists(tmpdir)) {
    -305  // Archive table in temp, maybe 
    left over from failed deletion,
    -306  // if not the cleaner will take 
    care of them.
    -307  for 

    [07/51] [partial] hbase-site git commit: Published site at 1d25b60831b8cc8f7ad5fd366f1867de5c20d2f3.

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/client/Scan.html
    --
    diff --git a/apidocs/org/apache/hadoop/hbase/client/Scan.html 
    b/apidocs/org/apache/hadoop/hbase/client/Scan.html
    index 806ad85..1ca681e 100644
    --- a/apidocs/org/apache/hadoop/hbase/client/Scan.html
    +++ b/apidocs/org/apache/hadoop/hbase/client/Scan.html
    @@ -97,7 +97,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">java.lang.Object
    +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">java.lang.Object
     
     
     org.apache.hadoop.hbase.client.Operation
    @@ -215,19 +215,19 @@ extends 
     
     
    -static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
    +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
     HBASE_CLIENT_SCANNER_ASYNC_PREFETCH
     Parameter name for client scanner sync/async prefetch 
    toggle.
     
     
     
    -static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
    +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
     SCAN_ATTRIBUTES_METRICS_DATA
     Deprecated.
     
     
     
    -static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
    +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
     SCAN_ATTRIBUTES_METRICS_ENABLE
     Deprecated.
     since 1.0.0. Use setScanMetricsEnabled(boolean)
    @@ -235,7 +235,7 @@ extends 
     
     
    -static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
    +static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
     SCAN_ATTRIBUTES_TABLE_NAME
     
     
    @@ -361,7 +361,7 @@ extends getFamilies()
     
     
    -http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
     title="class or interface in java.util">Mapbyte[],http://docs.oracle.com/javase/8/docs/api/java/util/NavigableSet.html?is-external=true;
     title="class or interface in 
    java.util">NavigableSetbyte[]
    +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
     title="class or interface in java.util">Mapbyte[],https://docs.oracle.com/javase/8/docs/api/java/util/NavigableSet.html?is-external=true;
     title="class or interface in 
    java.util">NavigableSetbyte[]
     getFamilyMap()
     Getting the familyMap
     
    @@ -371,7 +371,7 @@ extends getFilter()
     
     
    -http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
     title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
    +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
     title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     getFingerprint()
     Compile the table and column family (i.e.
     
    @@ -442,7 +442,7 @@ extends includeStopRow()
     
     
    -http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
     title="class or interface in java.lang">Boolean
    +https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
     title="class or interface in java.lang">Boolean
     isAsyncPrefetch()
     
     
    @@ -493,11 +493,11 @@ extends 
     
     Scan
    -setACL(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
     title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in 
    java.lang">String,org.apache.hadoop.hbase.security.access.Permissionperms)
    +setACL(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
     title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in 
    java.lang">String,org.apache.hadoop.hbase.security.access.Permissionperms)
     
     
     Scan
    -setACL(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">Stringuser,
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html 
    b/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html
    index e5fdac5..ad7c82a 100644
    --- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html
    +++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html
    @@ -33,299 +33,303 @@
     025import org.apache.hadoop.hbase.Cell;
     026import 
    org.apache.hadoop.hbase.CompareOperator;
     027import 
    org.apache.hadoop.hbase.PrivateCellUtil;
    -028import 
    org.apache.yetus.audience.InterfaceAudience;
    -029import 
    org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
    -030import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
    -031import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
    -032import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.CompareType;
    -033import 
    org.apache.hadoop.hbase.util.Bytes;
    -034
    -035import 
    org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
    -036/**
    -037 * This is a generic filter to be used to 
    filter by comparison.  It takes an
    -038 * operator (equal, greater, not equal, 
    etc) and a byte [] comparator.
    -039 * p
    -040 * To filter by row key, use {@link 
    RowFilter}.
    +028import 
    org.apache.hadoop.hbase.util.Bytes;
    +029import 
    org.apache.yetus.audience.InterfaceAudience;
    +030
    +031import 
    org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
    +032
    +033import 
    org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
    +034import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
    +035import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
    +036import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.CompareType;
    +037
    +038/**
    +039 * This is a generic filter to be used to 
    filter by comparison.  It takes an
    +040 * operator (equal, greater, not equal, 
    etc) and a byte [] comparator.
     041 * p
    -042 * To filter by column qualifier, use 
    {@link QualifierFilter}.
    +042 * To filter by row key, use {@link 
    RowFilter}.
     043 * p
    -044 * To filter by value, use {@link 
    SingleColumnValueFilter}.
    +044 * To filter by column family, use {@link 
    FamilyFilter}.
     045 * p
    -046 * These filters can be wrapped with 
    {@link SkipFilter} and {@link WhileMatchFilter}
    -047 * to add more control.
    -048 * p
    -049 * Multiple filters can be combined using 
    {@link FilterList}.
    -050 */
    -051@InterfaceAudience.Public
    -052public abstract class CompareFilter 
    extends FilterBase {
    -053  /**
    -054   * Comparison operators. For filters 
    only!
    -055   * Use {@link CompareOperator} 
    otherwise.
    -056   * It (intentionally) has at least the 
    below enums with same names.
    -057   * @deprecated  since 2.0.0. Will be 
    removed in 3.0.0. Use {@link CompareOperator} instead.
    -058   */
    -059  @Deprecated
    -060  @InterfaceAudience.Public
    -061  public enum CompareOp {
    -062/** less than */
    -063LESS,
    -064/** less than or equal to */
    -065LESS_OR_EQUAL,
    -066/** equals */
    -067EQUAL,
    -068/** not equal */
    -069NOT_EQUAL,
    -070/** greater than or equal to */
    -071GREATER_OR_EQUAL,
    -072/** greater than */
    -073GREATER,
    -074/** no operation */
    -075NO_OP,
    -076  }
    -077
    -078  protected CompareOperator op;
    -079  protected ByteArrayComparable 
    comparator;
    -080
    -081  /**
    -082   * Constructor.
    -083   * @param compareOp the compare op for 
    row matching
    -084   * @param comparator the comparator for 
    row matching
    -085   * @deprecated Since 2.0.0. Will be 
    removed in 3.0.0. Use other constructor.
    -086   */
    -087  @Deprecated
    -088  public CompareFilter(final CompareOp 
    compareOp,
    -089  final ByteArrayComparable 
    comparator) {
    -090
    this(CompareOperator.valueOf(compareOp.name()), comparator);
    -091  }
    -092
    -093  /**
    -094   * Constructor.
    -095   * @param op the compare op for row 
    matching
    -096   * @param comparator the comparator for 
    row matching
    -097   */
    -098  public CompareFilter(final 
    CompareOperator op,
    -099   final 
    ByteArrayComparable comparator) {
    -100this.op = op;
    -101this.comparator = comparator;
    -102  }
    -103
    -104  /**
    -105   * @return operator
    -106   * @deprecated  since 2.0.0. Will be 
    removed in 3.0.0. Use {@link #getCompareOperator()} instead.
    -107   */
    -108  @Deprecated
    -109  public CompareOp getOperator() {
    -110return 
    CompareOp.valueOf(op.name());
    -111  }
    -112
    -113  public CompareOperator 
    getCompareOperator() {
    -114return op;
    +046 * To filter by column qualifier, use 
    {@link QualifierFilter}.
    +047 * p
    +048 * To filter by value, use {@link 
    ValueFilter}.
    +049 * p
    +050 * These filters can be wrapped with 
    {@link SkipFilter} and {@link WhileMatchFilter}
    +051 * to add more control.
    +052 * p
    +053 * Multiple filters can be combined using 
    {@link FilterList}.
    +054 */
    +055@InterfaceAudience.Public
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
    index 802b925..a3e80ab 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
    @@ -73,229 +73,229 @@
     065import 
    java.util.concurrent.TimeoutException;
     066import 
    java.util.concurrent.atomic.AtomicBoolean;
     067import 
    java.util.concurrent.atomic.AtomicInteger;
    -068import 
    java.util.concurrent.atomic.AtomicLong;
    -069import 
    java.util.concurrent.atomic.LongAdder;
    -070import java.util.concurrent.locks.Lock;
    -071import 
    java.util.concurrent.locks.ReadWriteLock;
    -072import 
    java.util.concurrent.locks.ReentrantReadWriteLock;
    -073import java.util.function.Function;
    -074import 
    org.apache.hadoop.conf.Configuration;
    -075import org.apache.hadoop.fs.FileStatus;
    -076import org.apache.hadoop.fs.FileSystem;
    -077import 
    org.apache.hadoop.fs.LocatedFileStatus;
    -078import org.apache.hadoop.fs.Path;
    -079import org.apache.hadoop.hbase.Cell;
    -080import 
    org.apache.hadoop.hbase.CellBuilderType;
    -081import 
    org.apache.hadoop.hbase.CellComparator;
    -082import 
    org.apache.hadoop.hbase.CellComparatorImpl;
    -083import 
    org.apache.hadoop.hbase.CellScanner;
    -084import 
    org.apache.hadoop.hbase.CellUtil;
    -085import 
    org.apache.hadoop.hbase.CompareOperator;
    -086import 
    org.apache.hadoop.hbase.CompoundConfiguration;
    -087import 
    org.apache.hadoop.hbase.DoNotRetryIOException;
    -088import 
    org.apache.hadoop.hbase.DroppedSnapshotException;
    -089import 
    org.apache.hadoop.hbase.ExtendedCellBuilderFactory;
    -090import 
    org.apache.hadoop.hbase.HConstants;
    -091import 
    org.apache.hadoop.hbase.HConstants.OperationStatusCode;
    -092import 
    org.apache.hadoop.hbase.HDFSBlocksDistribution;
    -093import 
    org.apache.hadoop.hbase.HRegionInfo;
    -094import 
    org.apache.hadoop.hbase.KeyValue;
    -095import 
    org.apache.hadoop.hbase.KeyValueUtil;
    -096import 
    org.apache.hadoop.hbase.NamespaceDescriptor;
    -097import 
    org.apache.hadoop.hbase.NotServingRegionException;
    -098import 
    org.apache.hadoop.hbase.PrivateCellUtil;
    -099import 
    org.apache.hadoop.hbase.RegionTooBusyException;
    -100import 
    org.apache.hadoop.hbase.TableName;
    -101import org.apache.hadoop.hbase.Tag;
    -102import org.apache.hadoop.hbase.TagUtil;
    -103import 
    org.apache.hadoop.hbase.UnknownScannerException;
    -104import 
    org.apache.hadoop.hbase.client.Append;
    -105import 
    org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
    -106import 
    org.apache.hadoop.hbase.client.CompactionState;
    -107import 
    org.apache.hadoop.hbase.client.Delete;
    -108import 
    org.apache.hadoop.hbase.client.Durability;
    -109import 
    org.apache.hadoop.hbase.client.Get;
    -110import 
    org.apache.hadoop.hbase.client.Increment;
    -111import 
    org.apache.hadoop.hbase.client.IsolationLevel;
    -112import 
    org.apache.hadoop.hbase.client.Mutation;
    -113import 
    org.apache.hadoop.hbase.client.PackagePrivateFieldAccessor;
    -114import 
    org.apache.hadoop.hbase.client.Put;
    -115import 
    org.apache.hadoop.hbase.client.RegionInfo;
    -116import 
    org.apache.hadoop.hbase.client.RegionReplicaUtil;
    -117import 
    org.apache.hadoop.hbase.client.Result;
    -118import 
    org.apache.hadoop.hbase.client.RowMutations;
    -119import 
    org.apache.hadoop.hbase.client.Scan;
    -120import 
    org.apache.hadoop.hbase.client.TableDescriptor;
    -121import 
    org.apache.hadoop.hbase.client.TableDescriptorBuilder;
    -122import 
    org.apache.hadoop.hbase.conf.ConfigurationManager;
    -123import 
    org.apache.hadoop.hbase.conf.PropagatingConfigurationObserver;
    -124import 
    org.apache.hadoop.hbase.coprocessor.RegionObserver.MutationType;
    -125import 
    org.apache.hadoop.hbase.errorhandling.ForeignExceptionSnare;
    -126import 
    org.apache.hadoop.hbase.exceptions.FailedSanityCheckException;
    -127import 
    org.apache.hadoop.hbase.exceptions.TimeoutIOException;
    -128import 
    org.apache.hadoop.hbase.exceptions.UnknownProtocolException;
    -129import 
    org.apache.hadoop.hbase.filter.ByteArrayComparable;
    -130import 
    org.apache.hadoop.hbase.filter.FilterWrapper;
    -131import 
    org.apache.hadoop.hbase.filter.IncompatibleFilterException;
    -132import 
    org.apache.hadoop.hbase.io.HFileLink;
    -133import 
    org.apache.hadoop.hbase.io.HeapSize;
    -134import 
    org.apache.hadoop.hbase.io.TimeRange;
    -135import 
    org.apache.hadoop.hbase.io.hfile.HFile;
    -136import 
    org.apache.hadoop.hbase.ipc.CallerDisconnectedException;
    -137import 
    org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
    -138import 
    org.apache.hadoop.hbase.ipc.RpcCall;
    -139import 
    org.apache.hadoop.hbase.ipc.RpcServer;
    -140import 
    org.apache.hadoop.hbase.monitoring.MonitoredTask;
    -141import 
    org.apache.hadoop.hbase.monitoring.TaskMonitor;
    -142import 
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html 
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
    index bd13b53..802b925 100644
    --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
    +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
    @@ -900,7600 +900,7598 @@
     892if 
    (this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
     893  status.setStatus("Writing region 
    info on filesystem");
     894  fs.checkRegionInfoOnFilesystem();
    -895} else {
    -896  if (LOG.isDebugEnabled()) {
    -897LOG.debug("Skipping creation of 
    .regioninfo file for " + this.getRegionInfo());
    -898  }
    -899}
    -900
    -901// Initialize all the HStores
    -902status.setStatus("Initializing all 
    the Stores");
    -903long maxSeqId = 
    initializeStores(reporter, status);
    -904this.mvcc.advanceTo(maxSeqId);
    -905if 
    (ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
    -906  CollectionHStore stores = 
    this.stores.values();
    -907  try {
    -908// update the stores that we are 
    replaying
    -909
    stores.forEach(HStore::startReplayingFromWAL);
    -910// Recover any edits if 
    available.
    -911maxSeqId = Math.max(maxSeqId,
    -912  
    replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
    status));
    -913// Make sure mvcc is up to max.
    -914this.mvcc.advanceTo(maxSeqId);
    -915  } finally {
    -916// update the stores that we are 
    done replaying
    -917
    stores.forEach(HStore::stopReplayingFromWAL);
    -918  }
    -919}
    -920this.lastReplayedOpenRegionSeqId = 
    maxSeqId;
    +895}
    +896
    +897// Initialize all the HStores
    +898status.setStatus("Initializing all 
    the Stores");
    +899long maxSeqId = 
    initializeStores(reporter, status);
    +900this.mvcc.advanceTo(maxSeqId);
    +901if 
    (ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
    +902  CollectionHStore stores = 
    this.stores.values();
    +903  try {
    +904// update the stores that we are 
    replaying
    +905
    stores.forEach(HStore::startReplayingFromWAL);
    +906// Recover any edits if 
    available.
    +907maxSeqId = Math.max(maxSeqId,
    +908  
    replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
    status));
    +909// Make sure mvcc is up to max.
    +910this.mvcc.advanceTo(maxSeqId);
    +911  } finally {
    +912// update the stores that we are 
    done replaying
    +913
    stores.forEach(HStore::stopReplayingFromWAL);
    +914  }
    +915}
    +916this.lastReplayedOpenRegionSeqId = 
    maxSeqId;
    +917
    +918
    this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
    +919this.writestate.flushRequested = 
    false;
    +920this.writestate.compacting.set(0);
     921
    -922
    this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
    -923this.writestate.flushRequested = 
    false;
    -924this.writestate.compacting.set(0);
    -925
    -926if (this.writestate.writesEnabled) 
    {
    -927  // Remove temporary data left over 
    from old regions
    -928  status.setStatus("Cleaning up 
    temporary data from old regions");
    -929  fs.cleanupTempDir();
    -930}
    -931
    -932if (this.writestate.writesEnabled) 
    {
    -933  status.setStatus("Cleaning up 
    detritus from prior splits");
    -934  // Get rid of any splits or merges 
    that were lost in-progress.  Clean out
    -935  // these directories here on open.  
    We may be opening a region that was
    -936  // being split but we crashed in 
    the middle of it all.
    -937  fs.cleanupAnySplitDetritus();
    -938  fs.cleanupMergesDir();
    -939}
    -940
    -941// Initialize split policy
    -942this.splitPolicy = 
    RegionSplitPolicy.create(this, conf);
    -943
    -944// Initialize flush policy
    -945this.flushPolicy = 
    FlushPolicyFactory.create(this, conf);
    -946
    -947long lastFlushTime = 
    EnvironmentEdgeManager.currentTime();
    -948for (HStore store: stores.values()) 
    {
    -949  
    this.lastStoreFlushTimeMap.put(store, lastFlushTime);
    -950}
    -951
    -952// Use maximum of log sequenceid or 
    that which was found in stores
    -953// (particularly if no recovered 
    edits, seqid will be -1).
    -954long nextSeqid = maxSeqId;
    -955if (this.writestate.writesEnabled) 
    {
    -956  nextSeqid = 
    WALSplitter.writeRegionSequenceIdFile(this.fs.getFileSystem(),
    -957  this.fs.getRegionDir(), 
    nextSeqid, 1);
    -958} else {
    -959  nextSeqid++;
    -960}
    -961
    -962LOG.info("Onlined " + 
    this.getRegionInfo().getShortNameToLog() +
    -963  "; next sequenceid=" + 
    nextSeqid);
    +922if (this.writestate.writesEnabled) 
    {
    +923  // Remove temporary data left over 
    from old regions
    +924  status.setStatus("Cleaning up 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html 
    b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
    index 554f69f..e901330 100644
    --- a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
    +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
    @@ -132,13 +132,13 @@
     
     
     http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
     title="class or interface in java.util">ListProcedure?
    -MasterServices.getProcedures()
    -Get procedures
    -
    +HMaster.getProcedures()
     
     
     http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
     title="class or interface in java.util">ListProcedure?
    -HMaster.getProcedures()
    +MasterServices.getProcedures()
    +Get procedures
    +
     
     
     
    @@ -531,20 +531,20 @@
     Procedureproc)
     
     
    -boolean
    -PeerQueue.requireExclusiveLock(Procedure?proc)
    -
    -
     (package private) abstract boolean
     Queue.requireExclusiveLock(Procedure?proc)
     
    +
    +boolean
    +ServerQueue.requireExclusiveLock(Procedure?proc)
    +
     
     boolean
    -TableQueue.requireExclusiveLock(Procedure?proc)
    +PeerQueue.requireExclusiveLock(Procedure?proc)
     
     
     boolean
    -ServerQueue.requireExclusiveLock(Procedure?proc)
    +TableQueue.requireExclusiveLock(Procedure?proc)
     
     
     static byte[]
    @@ -920,44 +920,44 @@
     
     
     
    -protected Procedure
    -SimpleProcedureScheduler.dequeue()
    -
    -
     protected abstract Procedure
     AbstractProcedureScheduler.dequeue()
     Fetch one Procedure from the queue
      NOTE: this method is called with the sched lock held.
     
     
    -
    -protected Procedure[]
    -SequentialProcedure.doExecute(TEnvironmentenv)
    -
     
    +protected Procedure
    +SimpleProcedureScheduler.dequeue()
    +
    +
     protected ProcedureTEnvironment[]
     Procedure.doExecute(TEnvironmentenv)
     Internal method called by the ProcedureExecutor that starts 
    the user-level code execute().
     
     
    -
    -protected ProcedureTEnvironment[]
    -ProcedureExecutor.FailedProcedure.execute(TEnvironmentenv)
    -
     
     protected Procedure[]
    +SequentialProcedure.doExecute(TEnvironmentenv)
    +
    +
    +protected Procedure[]
     StateMachineProcedure.execute(TEnvironmentenv)
     
    +
    +protected ProcedureTEnvironment[]
    +ProcedureInMemoryChore.execute(TEnvironmentenv)
    +
     
    +protected ProcedureTEnvironment[]
    +ProcedureExecutor.FailedProcedure.execute(TEnvironmentenv)
    +
    +
     protected abstract ProcedureTEnvironment[]
     Procedure.execute(TEnvironmentenv)
     The main code of the procedure.
     
     
    -
    -protected ProcedureTEnvironment[]
    -ProcedureInMemoryChore.execute(TEnvironmentenv)
    -
     
     Procedure?
     LockedResource.getExclusiveLockOwnerProcedure()
    @@ -1115,14 +1115,14 @@
     
     
     void
    -SimpleProcedureScheduler.completionCleanup(Procedureproc)
    -
    -
    -void
     ProcedureScheduler.completionCleanup(Procedureproc)
     The procedure in execution completed.
     
     
    +
    +void
    +SimpleProcedureScheduler.completionCleanup(Procedureproc)
    +
     
     static 
    org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure
     ProcedureUtil.convertToProtoProcedure(Procedureproc)
    @@ -1135,17 +1135,17 @@
      Procedureprocedure)
     
     
    -protected void
    -SimpleProcedureScheduler.enqueue(Procedureprocedure,
    -   booleanaddFront)
    -
    -
     protected abstract void
     AbstractProcedureScheduler.enqueue(Procedureprocedure,
    booleanaddFront)
     Add the procedure to the queue.
     
     
    +
    +protected void
    +SimpleProcedureScheduler.enqueue(Procedureprocedure,
    +   booleanaddFront)
    +
     
     private void
     ProcedureExecutor.execCompletionCleanup(Procedureproc)
    @@ -1327,14 +1327,14 @@
     
     
     void
    -SimpleProcedureScheduler.yield(Procedureproc)
    -
    -
    -void
     ProcedureScheduler.yield(Procedureproc)
     The procedure can't run at the moment.
     
     
    +
    +void
    +SimpleProcedureScheduler.yield(Procedureproc)
    +
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html 
    b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html
    index a9a3870..5f35947 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html
    @@ -141,11 +141,11 @@
     
     
     ProcedureEvent?
    -MasterServices.getInitializedEvent()
    +HMaster.getInitializedEvent()
     
     
     ProcedureEvent?
    -HMaster.getInitializedEvent()
    +MasterServices.getInitializedEvent()
     
     
     ProcedureEvent?
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureExecutor.html
    --
    diff --git 
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html 
    b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html
    index 7244ce2..5f7ce59 100644
    --- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html
    +++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html
    @@ -114,15 +114,15 @@
     
     
     private PriorityFunction
    -SimpleRpcScheduler.priority
    +RpcExecutor.priority
     
     
     private PriorityFunction
    -RpcExecutor.priority
    +RpcExecutor.CallPriorityComparator.priority
     
     
     private PriorityFunction
    -RpcExecutor.CallPriorityComparator.priority
    +SimpleRpcScheduler.priority
     
     
     
    @@ -319,7 +319,7 @@
     
     
     RpcScheduler
    -RpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
    +FifoRpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
       PriorityFunctionpriority)
     Deprecated.
     
    @@ -333,18 +333,16 @@
     
     
     RpcScheduler
    -FifoRpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
    +RpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
       PriorityFunctionpriority)
     Deprecated.
     
     
     
     RpcScheduler
    -RpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
    +FifoRpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
       PriorityFunctionpriority,
    -  Abortableserver)
    -Constructs a RpcScheduler.
    -
    +  Abortableserver)
     
     
     RpcScheduler
    @@ -354,9 +352,11 @@
     
     
     RpcScheduler
    -FifoRpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
    +RpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
       PriorityFunctionpriority,
    -  Abortableserver)
    +  Abortableserver)
    +Constructs a RpcScheduler.
    +
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html 
    b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html
    index 4a25f5c..6d59fb7 100644
    --- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html
    +++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html
    @@ -123,14 +123,14 @@
     
     
     void
    -ServerCall.setCallBack(RpcCallbackcallback)
    -
    -
    -void
     RpcCallContext.setCallBack(RpcCallbackcallback)
     Sets a callback which has to be executed at the end of this 
    RPC call.
     
     
    +
    +void
    +ServerCall.setCallBack(RpcCallbackcallback)
    +
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html 
    b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html
    index fab4d7a..baa4e5e 100644
    --- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html
    +++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html
    @@ -131,32 +131,24 @@
     
     
     
    -private RpcControllerFactory
    -ConnectionImplementation.rpcControllerFactory
    -
    -
    -protected RpcControllerFactory
    -ClientScanner.rpcControllerFactory
    -
    -
     protected RpcControllerFactory
     RegionAdminServiceCallable.rpcControllerFactory
     
     
    -(package private) RpcControllerFactory
    -AsyncConnectionImpl.rpcControllerFactory
    +private RpcControllerFactory
    +ConnectionImplementation.rpcControllerFactory
     
     
    -private RpcControllerFactory
    -HTable.rpcControllerFactory
    +(package private) RpcControllerFactory
    +AsyncConnectionImpl.rpcControllerFactory
     
     
     private RpcControllerFactory
    -HBaseAdmin.rpcControllerFactory
    +HTable.rpcControllerFactory
     
     
     private RpcControllerFactory
    -SecureBulkLoadClient.rpcControllerFactory
    +RpcRetryingCallerWithReadReplicas.rpcControllerFactory
     
     
     protected RpcControllerFactory
    @@ -164,7 +156,15 @@
     
     
     private RpcControllerFactory
    -RpcRetryingCallerWithReadReplicas.rpcControllerFactory
    +HBaseAdmin.rpcControllerFactory
    +
    +
    +private RpcControllerFactory
    +SecureBulkLoadClient.rpcControllerFactory
    +
    +
    +protected RpcControllerFactory
    +ClientScanner.rpcControllerFactory
     
     
     (package private) RpcControllerFactory
    @@ -181,11 +181,11 @@
     
     
     RpcControllerFactory
    -ConnectionImplementation.getRpcControllerFactory()
    +ClusterConnection.getRpcControllerFactory()
     
     
     RpcControllerFactory
    -ClusterConnection.getRpcControllerFactory()
    +ConnectionImplementation.getRpcControllerFactory()
     
     
     private RpcControllerFactory
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcExecutor.Handler.html
    --
    diff --git 
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
    index 3628d68..bd2f966 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
    @@ -152,27 +152,27 @@ the order they are declared.
     
     
     PeerProcedureInterface.PeerOperationType
    -RefreshPeerProcedure.getPeerOperationType()
    +DisablePeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -DisablePeerProcedure.getPeerOperationType()
    +RemovePeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -UpdatePeerConfigProcedure.getPeerOperationType()
    +EnablePeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -AddPeerProcedure.getPeerOperationType()
    +RefreshPeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -EnablePeerProcedure.getPeerOperationType()
    +AddPeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -RemovePeerProcedure.getPeerOperationType()
    +UpdatePeerConfigProcedure.getPeerOperationType()
     
     
     private static PeerProcedureInterface.PeerOperationType
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
    index 269bc46..f7a6279 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
    @@ -125,11 +125,11 @@
     
     
     private ProcedurePrepareLatch
    -RecoverMetaProcedure.syncLatch
    +AbstractStateMachineTableProcedure.syncLatch
     
     
     private ProcedurePrepareLatch
    -AbstractStateMachineTableProcedure.syncLatch
    +RecoverMetaProcedure.syncLatch
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
    index 5e8085c..8b6ceb7 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
    @@ -104,14 +104,14 @@
     
     
     ServerProcedureInterface.ServerOperationType
    -ServerCrashProcedure.getServerOperationType()
    -
    -
    -ServerProcedureInterface.ServerOperationType
     ServerProcedureInterface.getServerOperationType()
     Given an operation type we can take decisions about what to 
    do with pending operations.
     
     
    +
    +ServerProcedureInterface.ServerOperationType
    +ServerCrashProcedure.getServerOperationType()
    +
     
     static ServerProcedureInterface.ServerOperationType
     ServerProcedureInterface.ServerOperationType.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">Stringname)
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
    index 046295e..e736f37 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
    @@ -112,19 +112,19 @@
     
     
     TableProcedureInterface.TableOperationType
    -MoveRegionProcedure.getTableOperationType()
    +UnassignProcedure.getTableOperationType()
     
     
     TableProcedureInterface.TableOperationType
    -GCMergedRegionsProcedure.getTableOperationType()
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html 
    b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html
    index abcb738..c7d05d1 100644
    --- a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html
    +++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html
    @@ -143,17 +143,17 @@
     
     
     void
    -NoOpDataBlockEncoder.saveMetadata(HFile.Writerwriter)
    +HFileDataBlockEncoderImpl.saveMetadata(HFile.Writerwriter)
     
     
     void
    -HFileDataBlockEncoder.saveMetadata(HFile.Writerwriter)
    -Save metadata in HFile which will be written to disk
    -
    +NoOpDataBlockEncoder.saveMetadata(HFile.Writerwriter)
     
     
     void
    -HFileDataBlockEncoderImpl.saveMetadata(HFile.Writerwriter)
    +HFileDataBlockEncoder.saveMetadata(HFile.Writerwriter)
    +Save metadata in HFile which will be written to disk
    +
     
     
     
    @@ -203,18 +203,18 @@
     
     
     
    -abstract void
    -BloomContext.addLastBloomKey(HFile.Writerwriter)
    -Adds the last bloom key to the HFile Writer as part of 
    StorefileWriter close.
    -
    +void
    +RowColBloomContext.addLastBloomKey(HFile.Writerwriter)
     
     
     void
     RowBloomContext.addLastBloomKey(HFile.Writerwriter)
     
     
    -void
    -RowColBloomContext.addLastBloomKey(HFile.Writerwriter)
    +abstract void
    +BloomContext.addLastBloomKey(HFile.Writerwriter)
    +Adds the last bloom key to the HFile Writer as part of 
    StorefileWriter close.
    +
     
     
     static BloomFilterWriter
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html 
    b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html
    index 274bfad..479b9d3 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html
    @@ -106,15 +106,15 @@
     
     
     
    -private HFileBlock.Writer
    -HFileBlockIndex.BlockIndexWriter.blockWriter
    -
    -
     protected HFileBlock.Writer
     HFileWriterImpl.blockWriter
     block writer
     
     
    +
    +private HFileBlock.Writer
    +HFileBlockIndex.BlockIndexWriter.blockWriter
    +
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html 
    b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html
    index b293c97..0c892c8 100644
    --- a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html
    +++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html
    @@ -136,15 +136,15 @@
     
     
     HFileContext
    -HFileBlockEncodingContext.getHFileContext()
    +HFileBlockDecodingContext.getHFileContext()
     
     
     HFileContext
    -HFileBlockDecodingContext.getHFileContext()
    +HFileBlockDefaultDecodingContext.getHFileContext()
     
     
     HFileContext
    -HFileBlockDefaultDecodingContext.getHFileContext()
    +HFileBlockEncodingContext.getHFileContext()
     
     
     HFileContext
    @@ -224,24 +224,24 @@
     
     
     private HFileContext
    -HFile.WriterFactory.fileContext
    -
    -
    -private HFileContext
     HFileBlock.fileContext
     Meta data that holds meta information on the 
    hfileblock.
     
     
    -
    +
     private HFileContext
     HFileBlock.Writer.fileContext
     Meta data that holds information about the hfileblock
     
     
    -
    +
     private HFileContext
     HFileBlock.FSReaderImpl.fileContext
     
    +
    +private HFileContext
    +HFile.WriterFactory.fileContext
    +
     
     private HFileContext
     HFileReaderImpl.hfileContext
    @@ -277,20 +277,20 @@
     
     
     HFileContext
    +HFileWriterImpl.getFileContext()
    +
    +
    +HFileContext
     HFile.Writer.getFileContext()
     Return the file context for the HFile this writer belongs 
    to
     
     
    -
    +
     HFileContext
     HFile.Reader.getFileContext()
     Return the file context of the HFile this reader belongs 
    to
     
     
    -
    -HFileContext
    -HFileWriterImpl.getFileContext()
    -
     
     HFileContext
     HFileReaderImpl.getFileContext()
    @@ -323,35 +323,35 @@
     
     
     HFileBlockDecodingContext
    -NoOpDataBlockEncoder.newDataBlockDecodingContext(HFileContextmeta)
    +HFileDataBlockEncoderImpl.newDataBlockDecodingContext(HFileContextfileContext)
     
     
     HFileBlockDecodingContext
    -HFileDataBlockEncoder.newDataBlockDecodingContext(HFileContextfileContext)
    -create a encoder specific decoding context for 
    reading.
    -
    +NoOpDataBlockEncoder.newDataBlockDecodingContext(HFileContextmeta)
     
     
     HFileBlockDecodingContext
    -HFileDataBlockEncoderImpl.newDataBlockDecodingContext(HFileContextfileContext)
    +HFileDataBlockEncoder.newDataBlockDecodingContext(HFileContextfileContext)
    +create a encoder specific decoding context for 
    reading.
    +
     
     
     HFileBlockEncodingContext
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
    index 3628d68..bd2f966 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
    @@ -152,27 +152,27 @@ the order they are declared.
     
     
     PeerProcedureInterface.PeerOperationType
    -RefreshPeerProcedure.getPeerOperationType()
    +DisablePeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -DisablePeerProcedure.getPeerOperationType()
    +RemovePeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -UpdatePeerConfigProcedure.getPeerOperationType()
    +EnablePeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -AddPeerProcedure.getPeerOperationType()
    +RefreshPeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -EnablePeerProcedure.getPeerOperationType()
    +AddPeerProcedure.getPeerOperationType()
     
     
     PeerProcedureInterface.PeerOperationType
    -RemovePeerProcedure.getPeerOperationType()
    +UpdatePeerConfigProcedure.getPeerOperationType()
     
     
     private static PeerProcedureInterface.PeerOperationType
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
    index 269bc46..f7a6279 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
    @@ -125,11 +125,11 @@
     
     
     private ProcedurePrepareLatch
    -RecoverMetaProcedure.syncLatch
    +AbstractStateMachineTableProcedure.syncLatch
     
     
     private ProcedurePrepareLatch
    -AbstractStateMachineTableProcedure.syncLatch
    +RecoverMetaProcedure.syncLatch
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
    index 5e8085c..8b6ceb7 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
    @@ -104,14 +104,14 @@
     
     
     ServerProcedureInterface.ServerOperationType
    -ServerCrashProcedure.getServerOperationType()
    -
    -
    -ServerProcedureInterface.ServerOperationType
     ServerProcedureInterface.getServerOperationType()
     Given an operation type we can take decisions about what to 
    do with pending operations.
     
     
    +
    +ServerProcedureInterface.ServerOperationType
    +ServerCrashProcedure.getServerOperationType()
    +
     
     static ServerProcedureInterface.ServerOperationType
     ServerProcedureInterface.ServerOperationType.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">Stringname)
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
    index 046295e..e736f37 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
    @@ -112,19 +112,19 @@
     
     
     TableProcedureInterface.TableOperationType
    -MoveRegionProcedure.getTableOperationType()
    +UnassignProcedure.getTableOperationType()
     
     
     TableProcedureInterface.TableOperationType
    -GCMergedRegionsProcedure.getTableOperationType()
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html 
    b/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html
    index 158a6aa..74a0687 100644
    --- a/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html
    +++ b/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html
    @@ -161,23 +161,23 @@
     
     
     Codec.Decoder
    -KeyValueCodec.getDecoder(ByteBuffbuf)
    +CellCodec.getDecoder(ByteBuffbuf)
     
     
     Codec.Decoder
    -CellCodecWithTags.getDecoder(ByteBuffbuf)
    +Codec.getDecoder(ByteBuffbuf)
     
     
     Codec.Decoder
    -Codec.getDecoder(ByteBuffbuf)
    +KeyValueCodec.getDecoder(ByteBuffbuf)
     
     
     Codec.Decoder
    -CellCodec.getDecoder(ByteBuffbuf)
    +KeyValueCodecWithTags.getDecoder(ByteBuffbuf)
     
     
     Codec.Decoder
    -KeyValueCodecWithTags.getDecoder(ByteBuffbuf)
    +CellCodecWithTags.getDecoder(ByteBuffbuf)
     
     
     Codec.Decoder
    @@ -259,20 +259,20 @@
     
     
     
    -protected ByteBuff
    -BufferedDataBlockEncoder.SeekerState.currentBuffer
    +private ByteBuff
    +RowIndexSeekerV1.currentBuffer
     
     
     protected ByteBuff
    -BufferedDataBlockEncoder.BufferedEncodedSeeker.currentBuffer
    +RowIndexSeekerV1.SeekerState.currentBuffer
     
     
    -private ByteBuff
    -RowIndexSeekerV1.currentBuffer
    +protected ByteBuff
    +BufferedDataBlockEncoder.SeekerState.currentBuffer
     
     
     protected ByteBuff
    -RowIndexSeekerV1.SeekerState.currentBuffer
    +BufferedDataBlockEncoder.BufferedEncodedSeeker.currentBuffer
     
     
     private ByteBuff
    @@ -295,23 +295,23 @@
     
     
     Cell
    -CopyKeyDataBlockEncoder.getFirstKeyCellInBlock(ByteBuffblock)
    +RowIndexCodecV1.getFirstKeyCellInBlock(ByteBuffblock)
     
     
     Cell
    -PrefixKeyDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
    +CopyKeyDataBlockEncoder.getFirstKeyCellInBlock(ByteBuffblock)
     
     
     Cell
    -FastDiffDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
    +DiffKeyDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
     
     
     Cell
    -DiffKeyDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
    +FastDiffDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
     
     
     Cell
    -RowIndexCodecV1.getFirstKeyCellInBlock(ByteBuffblock)
    +PrefixKeyDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
     
     
     void
    @@ -338,11 +338,11 @@
     
     
     void
    -BufferedDataBlockEncoder.BufferedEncodedSeeker.setCurrentBuffer(ByteBuffbuffer)
    +RowIndexSeekerV1.setCurrentBuffer(ByteBuffbuffer)
     
     
     void
    -RowIndexSeekerV1.setCurrentBuffer(ByteBuffbuffer)
    +BufferedDataBlockEncoder.BufferedEncodedSeeker.setCurrentBuffer(ByteBuffbuffer)
     
     
     
    @@ -498,21 +498,21 @@
     
     
     void
    -FileIOEngine.write(ByteBuffsrcBuffer,
    +ByteBufferIOEngine.write(ByteBuffsrcBuffer,
      longoffset)
     
     
     void
    +FileIOEngine.write(ByteBuffsrcBuffer,
    + longoffset)
    +
    +
    +void
     IOEngine.write(ByteBuffsrcBuffer,
      longoffset)
     Transfers the data from the given MultiByteBuffer to 
    IOEngine
     
     
    -
    -void
    -ByteBufferIOEngine.write(ByteBuffsrcBuffer,
    - longoffset)
    -
     
     void
     FileMmapEngine.write(ByteBuffsrcBuffer,
    @@ -812,15 +812,6 @@
      intindex)
     
     
    -MultiByteBuff
    -MultiByteBuff.put(intoffset,
    -   ByteBuffsrc,
    -   intsrcOffset,
    -   intlength)
    -Copies from a src MBB to this MBB.
    -
    -
    -
     abstract ByteBuff
     ByteBuff.put(intoffset,
    ByteBuffsrc,
    @@ -829,13 +820,22 @@
     Copies the contents from the src ByteBuff to this 
    ByteBuff.
     
     
    -
    +
     SingleByteBuff
     SingleByteBuff.put(intoffset,
    ByteBuffsrc,
    intsrcOffset,
    intlength)
     
    +
    +MultiByteBuff
    +MultiByteBuff.put(intoffset,
    +   ByteBuffsrc,
    +   intsrcOffset,
    +   intlength)
    +Copies from a src MBB to this MBB.
    +
    +
     
     static int
     ByteBuff.readCompressedInt(ByteBuffbuf)
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/package-tree.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/package-tree.html 
    b/devapidocs/org/apache/hadoop/hbase/package-tree.html
    index 96418c6..36170ca 100644
    --- a/devapidocs/org/apache/hadoop/hbase/package-tree.html
    +++ b/devapidocs/org/apache/hadoop/hbase/package-tree.html
    @@ -445,20 +445,20 @@
     
     java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
     title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
     title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
     title="class or interface in java.io">Serializable)
     
    -org.apache.hadoop.hbase.ClusterMetrics.Option
    -org.apache.hadoop.hbase.ProcedureState
     org.apache.hadoop.hbase.KeyValue.Type
    +org.apache.hadoop.hbase.Size.Unit
     org.apache.hadoop.hbase.CellBuilderType
     org.apache.hadoop.hbase.HealthChecker.HealthCheckerExitStatus
    +org.apache.hadoop.hbase.CompatibilitySingletonFactory.SingletonStorage
     org.apache.hadoop.hbase.MetaTableAccessor.QueryType
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
    index 6a0cad6..6fcf3a7 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
    @@ -158,7 +158,7 @@
     150// only one MasterCoprocessorHost 
    instance in the master process
     151boolean coprocessorsEnabled = 
    conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
     152  DEFAULT_COPROCESSORS_ENABLED);
    -153LOG.info("System coprocessor loading 
    is " + (coprocessorsEnabled ? "enabled" : "disabled"));
    +153LOG.trace("System coprocessor loading 
    is {}",  (coprocessorsEnabled ? "enabled" : "disabled"));
     154loadSystemCoprocessors(conf, 
    MASTER_COPROCESSOR_CONF_KEY);
     155  }
     156
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
    index 6a0cad6..6fcf3a7 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
    @@ -158,7 +158,7 @@
     150// only one MasterCoprocessorHost 
    instance in the master process
     151boolean coprocessorsEnabled = 
    conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
     152  DEFAULT_COPROCESSORS_ENABLED);
    -153LOG.info("System coprocessor loading 
    is " + (coprocessorsEnabled ? "enabled" : "disabled"));
    +153LOG.trace("System coprocessor loading 
    is {}",  (coprocessorsEnabled ? "enabled" : "disabled"));
     154loadSystemCoprocessors(conf, 
    MASTER_COPROCESSOR_CONF_KEY);
     155  }
     156
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
    index 6a0cad6..6fcf3a7 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
    @@ -158,7 +158,7 @@
     150// only one MasterCoprocessorHost 
    instance in the master process
     151boolean coprocessorsEnabled = 
    conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
     152  DEFAULT_COPROCESSORS_ENABLED);
    -153LOG.info("System coprocessor loading 
    is " + (coprocessorsEnabled ? "enabled" : "disabled"));
    +153LOG.trace("System coprocessor loading 
    is {}",  (coprocessorsEnabled ? "enabled" : "disabled"));
     154loadSystemCoprocessors(conf, 
    MASTER_COPROCESSOR_CONF_KEY);
     155  }
     156
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
    index 6a0cad6..6fcf3a7 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
    @@ -158,7 +158,7 @@
     150// only one MasterCoprocessorHost 
    instance in the master process
     151boolean coprocessorsEnabled = 
    conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
     152  DEFAULT_COPROCESSORS_ENABLED);
    -153LOG.info("System coprocessor loading 
    is " + (coprocessorsEnabled ? "enabled" : "disabled"));
    +153LOG.trace("System coprocessor loading 
    is {}",  (coprocessorsEnabled ? "enabled" : "disabled"));
     154loadSystemCoprocessors(conf, 
    MASTER_COPROCESSOR_CONF_KEY);
     155  }
     156
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html
    --
    diff --git 
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
    index 04bf42c..70c0a84 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
    @@ -271,527 +271,533 @@
     263return Flow.HAS_MORE_STATE;
     264  }
     265
    -266  @Override
    -267  protected void rollbackState(
    -268  final MasterProcedureEnv env,
    -269  final MergeTableRegionsState state) 
    throws IOException, InterruptedException {
    -270if (isTraceEnabled()) {
    -271  LOG.trace(this + " rollback state=" 
    + state);
    -272}
    -273
    -274try {
    -275  switch (state) {
    -276  case 
    MERGE_TABLE_REGIONS_POST_OPERATION:
    -277  case 
    MERGE_TABLE_REGIONS_OPEN_MERGED_REGION:
    -278  case 
    MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION:
    -279  case 
    MERGE_TABLE_REGIONS_UPDATE_META:
    -280String msg = this + " We are in 
    the " + state + " state."
    -281+ " It is complicated to 
    rollback the merge operation that region server is working on."
    -282+ " Rollback is not supported 
    and we should let the merge operation to complete";
    -283LOG.warn(msg);
    -284// PONR
    -285throw new 
    UnsupportedOperationException(this + " unhandled state=" + state);
    -286  case 
    MERGE_TABLE_REGIONS_PRE_MERGE_COMMIT_OPERATION:
    -287break;
    -288  case 
    MERGE_TABLE_REGIONS_CREATE_MERGED_REGION:
    -289cleanupMergedRegion(env);
    -290break;
    -291  case 
    MERGE_TABLE_REGIONS_CLOSE_REGIONS:
    -292
    rollbackCloseRegionsForMerge(env);
    +266  /**
    +267   * To rollback {@link 
    MergeTableRegionsProcedure}, two AssignProcedures are asynchronously
    +268   * submitted for each region to be 
    merged (rollback doesn't wait on the completion of the
    +269   * AssignProcedures) . This can be 
    improved by changing rollback() to support sub-procedures.
    +270   * See HBASE-19851 for details.
    +271   */
    +272  @Override
    +273  protected void rollbackState(
    +274  final MasterProcedureEnv env,
    +275  final MergeTableRegionsState state) 
    throws IOException, InterruptedException {
    +276if (isTraceEnabled()) {
    +277  LOG.trace(this + " rollback state=" 
    + state);
    +278}
    +279
    +280try {
    +281  switch (state) {
    +282  case 
    MERGE_TABLE_REGIONS_POST_OPERATION:
    +283  case 
    MERGE_TABLE_REGIONS_OPEN_MERGED_REGION:
    +284  case 
    MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION:
    +285  case 
    MERGE_TABLE_REGIONS_UPDATE_META:
    +286String msg = this + " We are in 
    the " + state + " state."
    +287+ " It is complicated to 
    rollback the merge operation that region server is working on."
    +288+ " Rollback is not supported 
    and we should let the merge operation to complete";
    +289LOG.warn(msg);
    +290// PONR
    +291throw new 
    UnsupportedOperationException(this + " unhandled state=" + state);
    +292  case 
    MERGE_TABLE_REGIONS_PRE_MERGE_COMMIT_OPERATION:
     293break;
    -294  case 
    MERGE_TABLE_REGIONS_PRE_MERGE_OPERATION:
    -295postRollBackMergeRegions(env);
    +294  case 
    MERGE_TABLE_REGIONS_CREATE_MERGED_REGION:
    +295cleanupMergedRegion(env);
     296break;
    -297  case MERGE_TABLE_REGIONS_PREPARE:
    -298break;
    -299  default:
    -300throw new 
    UnsupportedOperationException(this + " unhandled state=" + state);
    -301  }
    -302} catch (Exception e) {
    -303  // This will be retried. Unless 
    there is a bug in the code,
    -304  // this should be just a "temporary 
    error" (e.g. network down)
    -305  LOG.warn("Failed rollback attempt 
    step " + state + " for merging the regions "
    -306  + 
    RegionInfo.getShortNameToLog(regionsToMerge) + " in table " + getTableName(), 
    e);
    -307  throw e;
    -308}
    -309  }
    -310
    -311  /*
    -312   * Check whether we are in the state 
    that can be rollback
    -313   */
    -314  @Override
    -315  protected boolean 
    isRollbackSupported(final MergeTableRegionsState state) {
    -316switch (state) {
    -317case 
    MERGE_TABLE_REGIONS_POST_OPERATION:
    -318case 
    MERGE_TABLE_REGIONS_OPEN_MERGED_REGION:
    -319case 
    MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION:
    -320case 
    MERGE_TABLE_REGIONS_UPDATE_META:
    -321// It is not safe to rollback if 
    we reach to these states.
    -322return false;
    -323  default:
    -324break;
    -325}
    -326return true;
    -327  }
    -328
    -329  @Override
    -330  protected MergeTableRegionsState 
    getState(final int stateId) {
    -331return 
    MergeTableRegionsState.forNumber(stateId);
    -332  }
    -333
    -334  @Override
    -335  protected 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html
    --
    diff --git 
    a/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html 
    b/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html
    index ed294c6..4ba91bb 100644
    --- a/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html
    +++ b/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html
    @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -public class TestClientPushback
    +public class TestClientPushback
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     Test that we can actually send and use region metrics to 
    slowdown client writes
     
    @@ -131,26 +131,30 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     Field and Description
     
     
    +static HBaseClassTestRule
    +CLASS_RULE
    +
    +
     private static byte[]
     family
     
    -
    +
     private static long
     flushSizeBytes
     
    -
    +
     private static org.slf4j.Logger
     LOG
     
    -
    +
     private static byte[]
     qualifier
     
    -
    +
     private static 
    org.apache.hadoop.hbase.TableName
     tableName
     
    -
    +
     private static HBaseTestingUtility
     UTIL
     
    @@ -224,13 +228,22 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     Field Detail
    +
    +
    +
    +
    +
    +CLASS_RULE
    +public static finalHBaseClassTestRule CLASS_RULE
    +
    +
     
     
     
     
     
     LOG
    -private static finalorg.slf4j.Logger LOG
    +private static finalorg.slf4j.Logger LOG
     
     
     
    @@ -239,7 +252,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     UTIL
    -private static finalHBaseTestingUtility UTIL
    +private static finalHBaseTestingUtility UTIL
     
     
     
    @@ -248,7 +261,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     tableName
    -private static finalorg.apache.hadoop.hbase.TableName tableName
    +private static finalorg.apache.hadoop.hbase.TableName tableName
     
     
     
    @@ -257,7 +270,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     family
    -private static finalbyte[] family
    +private static finalbyte[] family
     
     
     
    @@ -266,7 +279,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     qualifier
    -private static finalbyte[] qualifier
    +private static finalbyte[] qualifier
     
     
     
    @@ -275,7 +288,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     flushSizeBytes
    -private static finallong flushSizeBytes
    +private static finallong flushSizeBytes
     
     See Also:
     Constant
     Field Values
    @@ -296,7 +309,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     TestClientPushback
    -publicTestClientPushback()
    +publicTestClientPushback()
     
     
     
    @@ -313,7 +326,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     setupCluster
    -public staticvoidsetupCluster()
    +public staticvoidsetupCluster()
      throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
     title="class or interface in java.lang">Exception
     
     Throws:
    @@ -327,7 +340,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     teardownCluster
    -public staticvoidteardownCluster()
    +public staticvoidteardownCluster()
     throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
     title="class or interface in java.lang">Exception
     
     Throws:
    @@ -341,7 +354,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     testClientTracksServerPushback
    -publicvoidtestClientTracksServerPushback()
    +publicvoidtestClientTracksServerPushback()
     throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
     title="class or interface in java.lang">Exception
     
     Throws:
    @@ -355,7 +368,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     testMutateRowStats
    -publicvoidtestMutateRowStats()
    +publicvoidtestMutateRowStats()
     throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
     title="class or interface in java.io">IOException
     
     Throws:
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
    --
    diff --git 
    a/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
     
    b/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
    index 38aa700..a0fc356 100644
    --- 
    a/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
    +++ 
    b/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
    @@ -132,7 +132,7 @@ var 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
    index e96923b..67775bf 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
    @@ -46,287 +46,272 @@
     038import 
    org.apache.hadoop.hbase.backup.impl.BackupManifest;
     039import 
    org.apache.hadoop.hbase.backup.impl.BackupSystemTable;
     040import 
    org.apache.hadoop.hbase.backup.util.BackupUtils;
    -041import 
    org.apache.yetus.audience.InterfaceAudience;
    -042import org.slf4j.Logger;
    -043import org.slf4j.LoggerFactory;
    -044import 
    org.apache.hadoop.hbase.client.Connection;
    -045import 
    org.apache.hadoop.hbase.client.ConnectionFactory;
    -046import 
    org.apache.hadoop.hbase.util.Pair;
    -047import org.apache.hadoop.util.Tool;
    +041import 
    org.apache.hadoop.hbase.client.Connection;
    +042import 
    org.apache.hadoop.hbase.client.ConnectionFactory;
    +043import 
    org.apache.hadoop.hbase.util.Pair;
    +044import org.apache.hadoop.util.Tool;
    +045import 
    org.apache.yetus.audience.InterfaceAudience;
    +046import org.slf4j.Logger;
    +047import org.slf4j.LoggerFactory;
     048
     049/**
     050 * MapReduce implementation of {@link 
    BackupMergeJob}
     051 * Must be initialized with configuration 
    of a backup destination cluster
     052 *
     053 */
    -054
    -055@InterfaceAudience.Private
    -056public class MapReduceBackupMergeJob 
    implements BackupMergeJob {
    -057  public static final Logger LOG = 
    LoggerFactory.getLogger(MapReduceBackupMergeJob.class);
    -058
    -059  protected Tool player;
    -060  protected Configuration conf;
    -061
    -062  public MapReduceBackupMergeJob() {
    -063  }
    -064
    -065  @Override
    -066  public Configuration getConf() {
    -067return conf;
    -068  }
    -069
    -070  @Override
    -071  public void setConf(Configuration conf) 
    {
    -072this.conf = conf;
    -073  }
    -074
    -075  @Override
    -076  public void run(String[] backupIds) 
    throws IOException {
    -077String bulkOutputConfKey;
    -078
    -079// TODO : run player on remote 
    cluster
    -080player = new 
    MapReduceHFileSplitterJob();
    -081bulkOutputConfKey = 
    MapReduceHFileSplitterJob.BULK_OUTPUT_CONF_KEY;
    -082// Player reads all files in 
    arbitrary directory structure and creates
    -083// a Map task for each file
    -084String bids = 
    StringUtils.join(backupIds, ",");
    -085
    -086if (LOG.isDebugEnabled()) {
    -087  LOG.debug("Merge backup images " + 
    bids);
    -088}
    -089
    -090ListPairTableName, 
    Path processedTableList = new ArrayListPairTableName, 
    Path();
    -091boolean finishedTables = false;
    -092Connection conn = 
    ConnectionFactory.createConnection(getConf());
    -093BackupSystemTable table = new 
    BackupSystemTable(conn);
    -094FileSystem fs = 
    FileSystem.get(getConf());
    -095
    -096try {
    -097
    -098  // Get exclusive lock on backup 
    system
    -099  
    table.startBackupExclusiveOperation();
    -100  // Start merge operation
    -101  
    table.startMergeOperation(backupIds);
    -102
    -103  // Select most recent backup id
    -104  String mergedBackupId = 
    findMostRecentBackupId(backupIds);
    -105
    -106  TableName[] tableNames = 
    getTableNamesInBackupImages(backupIds);
    -107  String backupRoot = null;
    -108
    -109  BackupInfo bInfo = 
    table.readBackupInfo(backupIds[0]);
    -110  backupRoot = 
    bInfo.getBackupRootDir();
    -111
    -112  for (int i = 0; i  
    tableNames.length; i++) {
    -113
    -114LOG.info("Merge backup images for 
    " + tableNames[i]);
    -115
    -116// Find input directories for 
    table
    -117
    -118Path[] dirPaths = 
    findInputDirectories(fs, backupRoot, tableNames[i], backupIds);
    -119String dirs = 
    StringUtils.join(dirPaths, ",");
    -120Path bulkOutputPath =
    -121
    BackupUtils.getBulkOutputDir(BackupUtils.getFileNameCompatibleString(tableNames[i]),
    -122  getConf(), false);
    -123// Delete content if exists
    -124if (fs.exists(bulkOutputPath)) 
    {
    -125  if (!fs.delete(bulkOutputPath, 
    true)) {
    -126LOG.warn("Can not delete: " + 
    bulkOutputPath);
    -127  }
    -128}
    -129Configuration conf = getConf();
    -130conf.set(bulkOutputConfKey, 
    bulkOutputPath.toString());
    -131String[] playerArgs = { dirs, 
    tableNames[i].getNameAsString() };
    -132
    -133int result = 0;
    -134
    -135player.setConf(getConf());
    -136result = 
    player.run(playerArgs);
    -137if (!succeeded(result)) {
    -138  throw new IOException("Can not 
    merge backup images for " + dirs
    -139  + " (check Hadoop/MR and 
    HBase logs). Player return code =" + result);
    -140}
    -141// Add to processed table list
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/testdevapidocs/overview-tree.html
    --
    diff --git a/testdevapidocs/overview-tree.html 
    b/testdevapidocs/overview-tree.html
    index 97691f1..ac787b0 100644
    --- a/testdevapidocs/overview-tree.html
    +++ b/testdevapidocs/overview-tree.html
    @@ -545,24 +545,6 @@
     org.apache.hadoop.hbase.io.TestImmutableBytesWritable
     org.apache.hadoop.hbase.TestKeyValue
     org.apache.hadoop.hbase.regionserver.TestKeyValueScanFixture
    -org.apache.hadoop.hbase.rest.model.TestModelBaseT
    -
    -org.apache.hadoop.hbase.rest.model.TestCellModel
    -org.apache.hadoop.hbase.rest.model.TestCellSetModel
    -org.apache.hadoop.hbase.rest.model.TestColumnSchemaModel
    -org.apache.hadoop.hbase.rest.model.TestNamespacesInstanceModel
    -org.apache.hadoop.hbase.rest.model.TestNamespacesModel
    -org.apache.hadoop.hbase.rest.model.TestRowModel
    -org.apache.hadoop.hbase.rest.model.TestScannerModel
    -org.apache.hadoop.hbase.rest.model.TestStorageClusterStatusModel
    -org.apache.hadoop.hbase.rest.model.TestStorageClusterVersionModel
    -org.apache.hadoop.hbase.rest.model.TestTableInfoModel
    -org.apache.hadoop.hbase.rest.model.TestTableListModel
    -org.apache.hadoop.hbase.rest.model.TestTableRegionModel
    -org.apache.hadoop.hbase.rest.model.TestTableSchemaModel
    -org.apache.hadoop.hbase.rest.model.TestVersionModel
    -
    -
     org.apache.hadoop.hbase.util.TestPoolMap.TestPoolType
     
     org.apache.hadoop.hbase.util.TestPoolMap.TestReusablePoolType
    @@ -970,7 +952,8 @@
     org.apache.hadoop.hbase.filter.TestFilterList.TransformFilter
     org.apache.hadoop.hbase.filter.TestInvocationRecordFilter.InvocationRecordFilter
     org.apache.hadoop.hbase.coprocessor.TestRegionObserverScannerOpenHook.NoDataFilter
    -org.apache.hadoop.hbase.regionserver.TestScannerHeartbeatMessages.SparseFilter
    +org.apache.hadoop.hbase.regionserver.TestScannerHeartbeatMessages.SparseCellFilter
    +org.apache.hadoop.hbase.regionserver.TestScannerHeartbeatMessages.SparseRowFilter
     org.apache.hadoop.hbase.regionserver.querymatcher.TestUserScanQueryMatcher.AlwaysIncludeAndSeekNextRowFilter
     org.apache.hadoop.hbase.regionserver.querymatcher.TestUserScanQueryMatcher.AlwaysIncludeFilter
     
    @@ -2915,6 +2898,24 @@
     org.apache.hadoop.hbase.mob.TestMobFileName
     org.apache.hadoop.hbase.regionserver.TestMobStoreCompaction
     org.apache.hadoop.hbase.regionserver.TestMobStoreScanner
    +org.apache.hadoop.hbase.rest.model.TestModelBaseT
    +
    +org.apache.hadoop.hbase.rest.model.TestCellModel
    +org.apache.hadoop.hbase.rest.model.TestCellSetModel
    +org.apache.hadoop.hbase.rest.model.TestColumnSchemaModel
    +org.apache.hadoop.hbase.rest.model.TestNamespacesInstanceModel
    +org.apache.hadoop.hbase.rest.model.TestNamespacesModel
    +org.apache.hadoop.hbase.rest.model.TestRowModel
    +org.apache.hadoop.hbase.rest.model.TestScannerModel
    +org.apache.hadoop.hbase.rest.model.TestStorageClusterStatusModel
    +org.apache.hadoop.hbase.rest.model.TestStorageClusterVersionModel
    +org.apache.hadoop.hbase.rest.model.TestTableInfoModel
    +org.apache.hadoop.hbase.rest.model.TestTableListModel
    +org.apache.hadoop.hbase.rest.model.TestTableRegionModel
    +org.apache.hadoop.hbase.rest.model.TestTableSchemaModel
    +org.apache.hadoop.hbase.rest.model.TestVersionModel
    +
    +
     org.apache.hadoop.hbase.master.procedure.TestModifyNamespaceProcedure
     org.apache.hadoop.hbase.TestMovedRegionsCleaner
     org.apache.hadoop.hbase.nio.TestMultiByteBuff
    @@ -3860,7 +3861,6 @@
     org.apache.hadoop.hbase.quotas.TestSnapshotQuotaObserverChore.IOThrowingRunnable
     org.apache.hadoop.hbase.master.TestSplitLogManager.Expr
     org.apache.hadoop.hbase.rest.TestTableScan.ClientSideCellSetModel.Listener
    -org.apache.hadoop.hbase.security.token.TestTokenAuthentication.AuthenticationServiceSecurityInfo
     org.apache.hadoop.hbase.TimestampTestBase.FlushCache
     org.apache.hadoop.hbase.testclassification.VerySlowMapReduceTests
     org.apache.hadoop.hbase.testclassification.VerySlowRegionServerTests
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
    --
    diff --git 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
     
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
    index 5c921e5..6009630 100644
    --- 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
    +++ 
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
    @@ -28,150 +28,148 @@
     020import static 
    org.junit.Assert.assertArrayEquals;
     021import static 
    org.junit.Assert.assertEquals;
     022import static 
    org.junit.Assert.assertFalse;
    -023
    -024import java.io.IOException;
    -025import 
    java.util.concurrent.CompletableFuture;
    -026import 
    java.util.concurrent.ExecutionException;
    -027
    -028import 
    org.apache.hadoop.hbase.testclassification.ClientTests;
    -029import 
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html 
    b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html
    index 6e2a4d4..3931484 100644
    --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html
    +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html
    @@ -182,7 +182,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     
    -private static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
    +(package private) static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
     REGION_TEMP_DIR
     Temporary subdirectory of the region directory used for 
    compaction output.
     
    @@ -708,7 +708,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     REGION_TEMP_DIR
    -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String REGION_TEMP_DIR
    +static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String REGION_TEMP_DIR
     Temporary subdirectory of the region directory used for 
    compaction output.
     
     See Also:
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
     
    b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
    index cd074b5..e5f8df7 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
    @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -private static class HRegionServer.CompactionChecker
    +private static class HRegionServer.CompactionChecker
     extends ScheduledChore
     
     
    @@ -233,7 +233,7 @@ extends 
     
     instance
    -private finalHRegionServer instance
    +private finalHRegionServer instance
     
     
     
    @@ -242,7 +242,7 @@ extends 
     
     majorCompactPriority
    -private finalint majorCompactPriority
    +private finalint majorCompactPriority
     
     
     
    @@ -251,7 +251,7 @@ extends 
     
     DEFAULT_PRIORITY
    -private static finalint DEFAULT_PRIORITY
    +private static finalint DEFAULT_PRIORITY
     
     See Also:
     Constant
     Field Values
    @@ -264,7 +264,7 @@ extends 
     
     iteration
    -privatelong iteration
    +privatelong iteration
     
     
     
    @@ -281,7 +281,7 @@ extends 
     
     CompactionChecker
    -CompactionChecker(HRegionServerh,
    +CompactionChecker(HRegionServerh,
       intsleepTime,
       Stoppablestopper)
     
    @@ -300,7 +300,7 @@ extends 
     
     chore
    -protectedvoidchore()
    +protectedvoidchore()
     Description copied from 
    class:ScheduledChore
     The task to execute on each scheduled execution of the 
    Chore
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
     
    b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
    index a387479..983e481 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
    @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -private static class HRegionServer.MovedRegionInfo
    +private static class HRegionServer.MovedRegionInfo
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     
     
    @@ -218,7 +218,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     serverName
    -private finalServerName serverName
    +private finalServerName serverName
     
     
     
    @@ -227,7 +227,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     seqNum
    -private finallong seqNum
    +private finallong seqNum
     
     
     
    @@ -236,7 +236,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     ts
    -private finallong ts
    +private finallong ts
     
     
     
    @@ -253,7 +253,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     MovedRegionInfo
    -publicMovedRegionInfo(ServerNameserverName,
    +publicMovedRegionInfo(ServerNameserverName,
    longcloseSeqNum)
     
     
    @@ -271,7 +271,7 @@ 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
    --
    diff --git 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
     
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
    index f1db5ca..d8515d7 100644
    --- 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
    +++ 
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
    @@ -32,813 +32,820 @@
     024import static org.junit.Assert.fail;
     025
     026import java.io.IOException;
    -027import java.net.SocketTimeoutException;
    -028import java.util.NavigableMap;
    -029import java.util.Random;
    -030import java.util.Set;
    -031import java.util.SortedSet;
    -032import 
    java.util.concurrent.ConcurrentSkipListMap;
    -033import 
    java.util.concurrent.ConcurrentSkipListSet;
    -034import 
    java.util.concurrent.ExecutionException;
    -035import java.util.concurrent.Executors;
    -036import java.util.concurrent.Future;
    -037import 
    java.util.concurrent.ScheduledExecutorService;
    -038import java.util.concurrent.TimeUnit;
    -039
    -040import 
    org.apache.hadoop.conf.Configuration;
    -041import 
    org.apache.hadoop.hbase.CategoryBasedTimeout;
    -042import 
    org.apache.hadoop.hbase.DoNotRetryIOException;
    -043import 
    org.apache.hadoop.hbase.HBaseTestingUtility;
    -044import 
    org.apache.hadoop.hbase.NotServingRegionException;
    -045import 
    org.apache.hadoop.hbase.ServerName;
    -046import 
    org.apache.hadoop.hbase.TableName;
    -047import 
    org.apache.hadoop.hbase.client.RegionInfo;
    -048import 
    org.apache.hadoop.hbase.client.RegionInfoBuilder;
    -049import 
    org.apache.hadoop.hbase.client.RetriesExhaustedException;
    -050import 
    org.apache.hadoop.hbase.exceptions.UnexpectedStateException;
    -051import 
    org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
    -052import 
    org.apache.hadoop.hbase.master.MasterServices;
    -053import 
    org.apache.hadoop.hbase.master.RegionState.State;
    -054import 
    org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants;
    -055import 
    org.apache.hadoop.hbase.master.procedure.ProcedureSyncWait;
    -056import 
    org.apache.hadoop.hbase.master.procedure.RSProcedureDispatcher;
    -057import 
    org.apache.hadoop.hbase.procedure2.Procedure;
    -058import 
    org.apache.hadoop.hbase.procedure2.ProcedureMetrics;
    -059import 
    org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility;
    -060import 
    org.apache.hadoop.hbase.procedure2.store.wal.WALProcedureStore;
    -061import 
    org.apache.hadoop.hbase.procedure2.util.StringUtils;
    -062import 
    org.apache.hadoop.hbase.regionserver.RegionServerAbortedException;
    -063import 
    org.apache.hadoop.hbase.regionserver.RegionServerStoppedException;
    -064import 
    org.apache.hadoop.hbase.testclassification.MasterTests;
    -065import 
    org.apache.hadoop.hbase.testclassification.MediumTests;
    -066import 
    org.apache.hadoop.hbase.util.Bytes;
    -067import 
    org.apache.hadoop.hbase.util.FSUtils;
    -068import 
    org.apache.hadoop.ipc.RemoteException;
    -069import org.junit.After;
    -070import org.junit.Before;
    -071import org.junit.Ignore;
    -072import org.junit.Rule;
    -073import org.junit.Test;
    -074import 
    org.junit.experimental.categories.Category;
    -075import 
    org.junit.rules.ExpectedException;
    -076import org.junit.rules.TestName;
    -077import org.junit.rules.TestRule;
    -078import org.slf4j.Logger;
    -079import org.slf4j.LoggerFactory;
    -080import 
    org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
    -081import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
    -082import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
    -083import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresRequest;
    -084import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresResponse;
    -085import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest;
    -086import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo;
    -087import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionResponse;
    -088import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState;
    -089import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition;
    -090import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode;
    -091import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionRequest;
    -092
    -093@Category({MasterTests.class, 
    MediumTests.class})
    -094public class TestAssignmentManager {
    -095  private static final Logger LOG = 
    LoggerFactory.getLogger(TestAssignmentManager.class);
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
    --
    diff --git 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
     
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
    index 232ef56..bc3a6d0 100644
    --- 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
    +++ 
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
    @@ -29,610 +29,626 @@
     021import static 
    org.junit.Assert.assertEquals;
     022import static 
    org.junit.Assert.assertFalse;
     023import static 
    org.junit.Assert.assertTrue;
    -024
    -025import java.io.ByteArrayOutputStream;
    -026import java.io.IOException;
    -027import java.math.BigDecimal;
    -028import java.nio.ByteBuffer;
    -029import java.util.ArrayList;
    -030import java.util.List;
    -031import java.util.NavigableMap;
    -032import java.util.TreeMap;
    -033import 
    org.apache.hadoop.hbase.testclassification.MiscTests;
    -034import 
    org.apache.hadoop.hbase.testclassification.SmallTests;
    -035import 
    org.apache.hadoop.hbase.util.Bytes;
    -036import org.junit.Assert;
    -037import org.junit.Test;
    -038import 
    org.junit.experimental.categories.Category;
    -039
    -040@Category({MiscTests.class, 
    SmallTests.class})
    -041public class TestCellUtil {
    -042  /**
    -043   * CellScannable used in test. Returns 
    a {@link TestCellScanner}
    -044   */
    -045  private static class TestCellScannable 
    implements CellScannable {
    -046private final int cellsCount;
    -047TestCellScannable(final int 
    cellsCount) {
    -048  this.cellsCount = cellsCount;
    -049}
    -050@Override
    -051public CellScanner cellScanner() {
    -052  return new 
    TestCellScanner(this.cellsCount);
    -053}
    -054  }
    -055
    -056  /**
    -057   * CellScanner used in test.
    -058   */
    -059  private static class TestCellScanner 
    implements CellScanner {
    -060private int count = 0;
    -061private Cell current = null;
    -062private final int cellsCount;
    -063
    -064TestCellScanner(final int cellsCount) 
    {
    -065  this.cellsCount = cellsCount;
    -066}
    -067
    -068@Override
    -069public Cell current() {
    -070  return this.current;
    -071}
    -072
    -073@Override
    -074public boolean advance() throws 
    IOException {
    -075  if (this.count  cellsCount) {
    -076this.current = new 
    TestCell(this.count);
    -077this.count++;
    -078return true;
    -079  }
    -080  return false;
    -081}
    -082  }
    -083
    -084  /**
    -085   * Cell used in test. Has row only.
    -086   */
    -087  private static class TestCell 
    implements Cell {
    -088private final byte [] row;
    -089
    -090TestCell(final int i) {
    -091  this.row = Bytes.toBytes(i);
    -092}
    -093
    -094@Override
    -095public byte[] getRowArray() {
    -096  return this.row;
    -097}
    -098
    -099@Override
    -100public int getRowOffset() {
    -101  return 0;
    -102}
    -103
    -104@Override
    -105public short getRowLength() {
    -106  return (short)this.row.length;
    -107}
    -108
    -109@Override
    -110public byte[] getFamilyArray() {
    -111  // TODO Auto-generated method 
    stub
    -112  return null;
    -113}
    -114
    -115@Override
    -116public int getFamilyOffset() {
    -117  // TODO Auto-generated method 
    stub
    -118  return 0;
    -119}
    -120
    -121@Override
    -122public byte getFamilyLength() {
    -123  // TODO Auto-generated method 
    stub
    -124  return 0;
    -125}
    -126
    -127@Override
    -128public byte[] getQualifierArray() {
    -129  // TODO Auto-generated method 
    stub
    -130  return null;
    -131}
    -132
    -133@Override
    -134public int getQualifierOffset() {
    -135  // TODO Auto-generated method 
    stub
    -136  return 0;
    -137}
    -138
    -139@Override
    -140public int getQualifierLength() {
    -141  // TODO Auto-generated method 
    stub
    -142  return 0;
    -143}
    -144
    -145@Override
    -146public long getTimestamp() {
    -147  // TODO Auto-generated method 
    stub
    -148  return 0;
    -149}
    -150
    -151@Override
    -152public byte getTypeByte() {
    -153  // TODO Auto-generated method 
    stub
    -154  return 0;
    -155}
    -156
    -157@Override
    -158public byte[] getValueArray() {
    -159  // TODO Auto-generated method 
    stub
    -160  return null;
    -161}
    -162
    -163@Override
    -164public int getValueOffset() {
    -165  // TODO Auto-generated method 
    stub
    -166  return 0;
    -167}
    -168
    -169@Override
    -170public int getValueLength() {
    -171  // TODO Auto-generated method 
    stub
    -172  return 0;
    -173}
    -174
    -175@Override
    -176public byte[] getTagsArray() {
    -177  // TODO Auto-generated method 
    stub
    -178  return null;
    -179}
    -180
    -181@Override
    -182public int getTagsOffset() {
    -183  // TODO Auto-generated method 
    stub
    -184  return 0;
    -185}
    -186
    -187@Override
    -188public long getSequenceId() {
    -189   

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html
    --
    diff --git 
    a/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html 
    b/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html
    index 5f0e47f..8a340d3 100644
    --- 
    a/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html
    +++ 
    b/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html
    @@ -18,7 +18,7 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":10,"i1":9,"i2":10,"i3":10,"i4":9,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10};
    +var methods = 
    {"i0":10,"i1":10,"i2":9,"i3":10,"i4":10,"i5":9,"i6":10,"i7":9,"i8":10,"i9":10,"i10":10,"i11":10};
     var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
    @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -public class TestDurability
    +public class TestDurability
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     Tests for WAL write durability
     
    @@ -155,14 +155,18 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     FS
     
     
    +org.junit.rules.TestName
    +name
    +
    +
     private static byte[]
     ROW
     
    -
    +
     private static HBaseTestingUtility
     TEST_UTIL
     
    -
    +
     http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
     walProvider
     
    @@ -200,48 +204,57 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     private 
    org.apache.hadoop.hbase.regionserver.HRegion
    -createHRegion(byte[]tableName,
    - http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">StringcallingMethod,
    - org.apache.hadoop.hbase.wal.WALlog,
    +createHRegion(org.apache.hadoop.hbase.client.TableDescriptortd,
    + org.apache.hadoop.hbase.client.RegionInfoinfo,
    + http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">Stringdir,
    + org.apache.hadoop.hbase.wal.WALwal,
      
    org.apache.hadoop.hbase.client.Durabilitydurability)
     
     
    +private 
    org.apache.hadoop.hbase.regionserver.HRegion
    +createHRegion(org.apache.hadoop.hbase.wal.WALFactorywals,
    + 
    org.apache.hadoop.hbase.client.Durabilitydurability)
    +
    +
     static http://docs.oracle.com/javase/8/docs/api/java/lang/Iterable.html?is-external=true;
     title="class or interface in java.lang">Iterablehttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object[]
     data()
     
    -
    +
     private 
    org.apache.hadoop.hbase.client.Put
     newPut(org.apache.hadoop.hbase.client.Durabilitydurability)
     
    -
    +
     void
     setUp()
     
    -
    +
     static void
     setUpBeforeClass()
     
    -
    +
     void
     tearDown()
     
    -
    +
     static void
     tearDownAfterClass()
     
    -
    +
     void
     testDurability()
     
    -
    +
     void
     testIncrement()
     
    -
    +
     void
    -testIncrementWithReturnResultsSetToFalse()
    +testIncrementWithReturnResultsSetToFalse()
    +Test when returnResults set to false in increment it should 
    not return the result instead it
    + resturn null.
    +
     
    -
    +
     private void
     verifyWALCount(org.apache.hadoop.hbase.wal.WALFactorywals,
       org.apache.hadoop.hbase.wal.WALlog,
    @@ -275,7 +288,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     TEST_UTIL
    -private static finalHBaseTestingUtility TEST_UTIL
    +private static finalHBaseTestingUtility TEST_UTIL
     
     
     
    @@ -284,7 +297,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     FS
    -private staticorg.apache.hadoop.fs.FileSystem FS
    +private staticorg.apache.hadoop.fs.FileSystem FS
     
     
     
    @@ -293,7 +306,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     CLUSTER
    -private staticorg.apache.hadoop.hdfs.MiniDFSCluster CLUSTER
    +private staticorg.apache.hadoop.hdfs.MiniDFSCluster CLUSTER
     
     
     
    @@ -302,7 +315,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     CONF
    -private staticorg.apache.hadoop.conf.Configuration CONF
    +private staticorg.apache.hadoop.conf.Configuration CONF
     
     
     
    @@ -311,7 +324,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     DIR
    -private staticorg.apache.hadoop.fs.Path DIR
    +private staticorg.apache.hadoop.fs.Path DIR
     
     
     
    @@ -320,7 +333,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     FAMILY
    -private staticbyte[] FAMILY
    +private staticbyte[] FAMILY
     
     
     
    @@ -329,7 +342,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     ROW
    -private staticbyte[] ROW
    +private staticbyte[] ROW
     
     
     
    @@ -338,16 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
     
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
    index 3ab8444..28dc97a 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
    @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
     
     
     PrevClass
    -NextClass
    +NextClass
     
     
     Frames
    @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -private static class MasterProcedureScheduler.FairQueueT
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
     title="class or interface in java.lang">ComparableT
    +private static class MasterProcedureScheduler.FairQueueT
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
     title="class or interface in java.lang">ComparableT
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     
     
    @@ -244,7 +244,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     quantum
    -private finalint quantum
    +private finalint quantum
     
     
     
    @@ -253,7 +253,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     currentQueue
    -privateMasterProcedureScheduler.QueueT extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
     title="class or interface in java.lang">ComparableT currentQueue
    +privateMasterProcedureScheduler.QueueT extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
     title="class or interface in java.lang">ComparableT currentQueue
     
     
     
    @@ -262,7 +262,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     queueHead
    -privateMasterProcedureScheduler.QueueT extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
     title="class or interface in java.lang">ComparableT queueHead
    +privateMasterProcedureScheduler.QueueT extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
     title="class or interface in java.lang">ComparableT queueHead
     
     
     
    @@ -271,7 +271,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     currentQuantum
    -privateint currentQuantum
    +privateint currentQuantum
     
     
     
    @@ -280,7 +280,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     size
    -privateint size
    +privateint size
     
     
     
    @@ -297,7 +297,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     FairQueue
    -publicFairQueue()
    +publicFairQueue()
     
     
     
    @@ -306,7 +306,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     FairQueue
    -publicFairQueue(intquantum)
    +publicFairQueue(intquantum)
     
     
     
    @@ -323,7 +323,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     hasRunnables
    -publicbooleanhasRunnables()
    +publicbooleanhasRunnables()
     
     
     
    @@ -332,7 +332,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     add
    -publicvoidadd(MasterProcedureScheduler.QueueTqueue)
    +publicvoidadd(MasterProcedureScheduler.QueueTqueue)
     
     
     
    @@ -341,7 +341,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     remove
    -publicvoidremove(MasterProcedureScheduler.QueueTqueue)
    +publicvoidremove(MasterProcedureScheduler.QueueTqueue)
     
     
     
    @@ -350,7 +350,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     poll
    -publicMasterProcedureScheduler.QueueTpoll()
    +publicMasterProcedureScheduler.QueueTpoll()
     
     
     
    @@ -359,7 +359,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     nextQueue
    -privatebooleannextQueue()
    +privatebooleannextQueue()
     
     
     
    @@ -368,7 +368,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     setNextQueue
    -privatevoidsetNextQueue(MasterProcedureScheduler.QueueTqueue)
    +privatevoidsetNextQueue(MasterProcedureScheduler.QueueTqueue)
     
     
     
    @@ -377,7 +377,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     calculateQuantum
    -privateintcalculateQuantum(MasterProcedureScheduler.Queuequeue)
    +privateintcalculateQuantum(MasterProcedureScheduler.Queuequeue)
     
     
     
    @@ -409,7 +409,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     PrevClass
    -NextClass
    +NextClass
     
     
     Frames
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.PeerQueue.html
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html 
    b/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
    index 9e7d3ae..c7b483d 100644
    --- a/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
    +++ b/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
    @@ -18,7 +18,7 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":9,"i25":9,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109"
     
    :10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10};
    +var methods = 
    {"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":9,"i25":9,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109"
     
    :10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10};
     var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
    @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
     
     
     @InterfaceAudience.LimitedPrivate(value="Configuration")
    -public class AccessController
    +public class AccessController
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     implements MasterCoprocessor, RegionCoprocessor, RegionServerCoprocessor, 
    org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.Interface,
     MasterObserver, 
    RegionObserver, 
    RegionServerObserver, EndpointObserver, 
    BulkLoadObserver
     Provides basic authorization checks for data access and 
    administrative
    @@ -959,24 +959,18 @@ implements 
     void
    -preGetClusterStatus(ObserverContextMasterCoprocessorEnvironmentctx)
    -Called before get cluster status.
    -
    -
    -
    -void
     preGetLocks(ObserverContextMasterCoprocessorEnvironmentctx)
     Called before a getLocks request has been processed.
     
     
    -
    +
     void
     preGetNamespaceDescriptor(ObserverContextMasterCoprocessorEnvironmentctx,
      http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">Stringnamespace)
     Called before a getNamespaceDescriptor request has been 
    processed.
     
     
    -
    +
     void
     preGetOp(ObserverContextRegionCoprocessorEnvironmentc,
     Getget,
    @@ -984,20 +978,20 @@ implements Called before the client performs a Get
     
     
    -
    +
     

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
    index 8373f27..80df615 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
    @@ -6,294 +6,293 @@
     
     
     
    -001
    -002/**
    -003 * Copyright The Apache Software 
    Foundation
    -004 *
    -005 * Licensed to the Apache Software 
    Foundation (ASF) under one or more
    -006 * contributor license agreements. See 
    the NOTICE file distributed with this
    -007 * work for additional information 
    regarding copyright ownership. The ASF
    -008 * licenses this file to you under the 
    Apache License, Version 2.0 (the
    -009 * "License"); you may not use this file 
    except in compliance with the License.
    -010 * You may obtain a copy of the License 
    at
    -011 *
    -012 * 
    http://www.apache.org/licenses/LICENSE-2.0
    -013 *
    -014 * Unless required by applicable law or 
    agreed to in writing, software
    -015 * distributed under the License is 
    distributed on an "AS IS" BASIS, WITHOUT
    -016 * WARRANTIES OR CONDITIONS OF ANY KIND, 
    either express or implied. See the
    -017 * License for the specific language 
    governing permissions and limitations
    -018 * under the License.
    -019 */
    -020
    -021package 
    org.apache.hadoop.hbase.io.hfile;
    -022
    -023import java.io.IOException;
    -024import java.net.InetSocketAddress;
    -025import java.nio.ByteBuffer;
    -026import java.util.ArrayList;
    -027import java.util.Iterator;
    -028import java.util.List;
    -029import 
    java.util.NoSuchElementException;
    -030import 
    java.util.concurrent.ExecutionException;
    -031
    -032import 
    org.apache.hadoop.conf.Configuration;
    -033import 
    org.apache.hadoop.hbase.HConstants;
    -034import 
    org.apache.hadoop.hbase.io.hfile.Cacheable.MemoryType;
    -035import 
    org.apache.hadoop.hbase.nio.ByteBuff;
    -036import 
    org.apache.hadoop.hbase.nio.SingleByteBuff;
    -037import 
    org.apache.hadoop.hbase.trace.TraceUtil;
    -038import 
    org.apache.hadoop.hbase.util.Addressing;
    -039import 
    org.apache.htrace.core.TraceScope;
    -040import 
    org.apache.yetus.audience.InterfaceAudience;
    -041import org.slf4j.Logger;
    -042import org.slf4j.LoggerFactory;
    -043
    -044import net.spy.memcached.CachedData;
    -045import 
    net.spy.memcached.ConnectionFactoryBuilder;
    -046import net.spy.memcached.FailureMode;
    -047import 
    net.spy.memcached.MemcachedClient;
    -048import 
    net.spy.memcached.transcoders.Transcoder;
    -049
    -050/**
    -051 * Class to store blocks into 
    memcached.
    -052 * This should only be used on a cluster 
    of Memcached daemons that are tuned well and have a
    -053 * good network connection to the HBase 
    regionservers. Any other use will likely slow down HBase
    -054 * greatly.
    -055 */
    -056@InterfaceAudience.Private
    -057public class MemcachedBlockCache 
    implements BlockCache {
    -058  private static final Logger LOG = 
    LoggerFactory.getLogger(MemcachedBlockCache.class.getName());
    -059
    -060  // Some memcache versions won't take 
    more than 1024 * 1024. So set the limit below
    -061  // that just in case this client is 
    used with those versions.
    -062  public static final int MAX_SIZE = 1020 
    * 1024;
    -063
    -064  // Config key for what memcached 
    servers to use.
    -065  // They should be specified in a comma 
    sperated list with ports.
    -066  // like:
    -067  //
    -068  // host1:11211,host3:8080,host4:11211
    -069  public static final String 
    MEMCACHED_CONFIG_KEY = "hbase.cache.memcached.servers";
    -070  public static final String 
    MEMCACHED_TIMEOUT_KEY = "hbase.cache.memcached.timeout";
    -071  public static final String 
    MEMCACHED_OPTIMEOUT_KEY = "hbase.cache.memcached.optimeout";
    -072  public static final String 
    MEMCACHED_OPTIMIZE_KEY = "hbase.cache.memcached.spy.optimze";
    -073  public static final long 
    MEMCACHED_DEFAULT_TIMEOUT = 500;
    -074  public static final boolean 
    MEMCACHED_OPTIMIZE_DEFAULT = false;
    -075
    -076  private final MemcachedClient client;
    -077  private final HFileBlockTranscoder tc = 
    new HFileBlockTranscoder();
    -078  private final CacheStats cacheStats = 
    new CacheStats("MemcachedBlockCache");
    -079
    -080  public 
    MemcachedBlockCache(Configuration c) throws IOException {
    -081LOG.info("Creating 
    MemcachedBlockCache");
    -082
    -083long opTimeout = 
    c.getLong(MEMCACHED_OPTIMEOUT_KEY, MEMCACHED_DEFAULT_TIMEOUT);
    -084long queueTimeout = 
    c.getLong(MEMCACHED_TIMEOUT_KEY, opTimeout + MEMCACHED_DEFAULT_TIMEOUT);
    -085boolean optimize = 
    c.getBoolean(MEMCACHED_OPTIMIZE_KEY, MEMCACHED_OPTIMIZE_DEFAULT);
    -086
    -087ConnectionFactoryBuilder builder = 
    new ConnectionFactoryBuilder()
    -088.setOpTimeout(opTimeout)
    -089
    .setOpQueueMaxBlockTime(queueTimeout) 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html 
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html
    index a591b29..030d8dd 100644
    --- a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html
    +++ b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html
    @@ -120,6 +120,6 @@
     
     
     
    -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
    +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html 
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html
    index 588d268..a9e5ff5 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html
    @@ -120,6 +120,6 @@
     
     
     
    -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
    +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html 
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html
    index 5c452b2..cf51754 100644
    --- a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html
    +++ b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html
    @@ -120,6 +120,6 @@
     
     
     
    -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
    +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html 
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html
    index 5738603..a5e9992 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html
    @@ -120,6 +120,6 @@
     
     
     
    -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
    +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
     
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
    index 99e8e24..f29b6e6 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
    @@ -120,6 +120,6 @@
     
     
     
    -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
    +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
     
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
    index 7971d8d..91a5c17 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
    @@ -172,6 +172,6 @@ the order they are declared.
     
     
     
    -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
    +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
    reserved.
     
     
    
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
    --
    diff --git 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
     
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
    index bbd91b8..4f76302 100644
    --- 
    a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
    +++ 
    b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
    @@ -56,1641 +56,1753 @@
     048import 
    java.util.concurrent.atomic.AtomicBoolean;
     049import 
    java.util.concurrent.atomic.AtomicInteger;
     050import 
    java.util.concurrent.atomic.AtomicLong;
    -051
    -052import 
    org.apache.hadoop.conf.Configuration;
    -053import 
    org.apache.hadoop.hbase.CallQueueTooBigException;
    -054import 
    org.apache.hadoop.hbase.CategoryBasedTimeout;
    -055import org.apache.hadoop.hbase.Cell;
    -056import 
    org.apache.hadoop.hbase.HConstants;
    -057import 
    org.apache.hadoop.hbase.HRegionInfo;
    -058import 
    org.apache.hadoop.hbase.HRegionLocation;
    -059import 
    org.apache.hadoop.hbase.RegionLocations;
    -060import 
    org.apache.hadoop.hbase.ServerName;
    -061import 
    org.apache.hadoop.hbase.TableName;
    -062import 
    org.apache.hadoop.hbase.client.AsyncProcessTask.ListRowAccess;
    -063import 
    org.apache.hadoop.hbase.client.AsyncProcessTask.SubmittedRows;
    -064import 
    org.apache.hadoop.hbase.client.backoff.ClientBackoffPolicy;
    -065import 
    org.apache.hadoop.hbase.client.backoff.ServerStatistics;
    -066import 
    org.apache.hadoop.hbase.client.coprocessor.Batch;
    -067import 
    org.apache.hadoop.hbase.ipc.RpcControllerFactory;
    -068import 
    org.apache.hadoop.hbase.testclassification.ClientTests;
    -069import 
    org.apache.hadoop.hbase.testclassification.MediumTests;
    -070import 
    org.apache.hadoop.hbase.util.Bytes;
    -071import 
    org.apache.hadoop.hbase.util.Threads;
    -072import org.junit.Assert;
    -073import org.junit.BeforeClass;
    -074import org.junit.Ignore;
    -075import org.junit.Rule;
    -076import org.junit.Test;
    -077import 
    org.junit.experimental.categories.Category;
    -078import org.junit.rules.TestRule;
    -079import org.mockito.Mockito;
    -080import org.slf4j.Logger;
    -081import org.slf4j.LoggerFactory;
    -082
    -083@Category({ClientTests.class, 
    MediumTests.class})
    -084public class TestAsyncProcess {
    -085  @Rule public final TestRule timeout = 
    CategoryBasedTimeout.builder().withTimeout(this.getClass()).
    -086  
    withLookingForStuckThread(true).build();
    -087  private static final Logger LOG = 
    LoggerFactory.getLogger(TestAsyncProcess.class);
    -088  private static final TableName 
    DUMMY_TABLE =
    -089  TableName.valueOf("DUMMY_TABLE");
    -090  private static final byte[] 
    DUMMY_BYTES_1 = Bytes.toBytes("DUMMY_BYTES_1");
    -091  private static final byte[] 
    DUMMY_BYTES_2 = Bytes.toBytes("DUMMY_BYTES_2");
    -092  private static final byte[] 
    DUMMY_BYTES_3 = Bytes.toBytes("DUMMY_BYTES_3");
    -093  private static final byte[] FAILS = 
    Bytes.toBytes("FAILS");
    -094  private static final Configuration CONF 
    = new Configuration();
    -095  private static final 
    ConnectionConfiguration CONNECTION_CONFIG =
    -096  new 
    ConnectionConfiguration(CONF);
    -097  private static final ServerName sn = 
    ServerName.valueOf("s1,1,1");
    -098  private static final ServerName sn2 = 
    ServerName.valueOf("s2,2,2");
    -099  private static final ServerName sn3 = 
    ServerName.valueOf("s3,3,3");
    -100  private static final HRegionInfo hri1 
    =
    -101  new HRegionInfo(DUMMY_TABLE, 
    DUMMY_BYTES_1, DUMMY_BYTES_2, false, 1);
    -102  private static final HRegionInfo hri2 
    =
    -103  new HRegionInfo(DUMMY_TABLE, 
    DUMMY_BYTES_2, HConstants.EMPTY_END_ROW, false, 2);
    -104  private static final HRegionInfo hri3 
    =
    -105  new HRegionInfo(DUMMY_TABLE, 
    DUMMY_BYTES_3, HConstants.EMPTY_END_ROW, false, 3);
    -106  private static final HRegionLocation 
    loc1 = new HRegionLocation(hri1, sn);
    -107  private static final HRegionLocation 
    loc2 = new HRegionLocation(hri2, sn);
    -108  private static final HRegionLocation 
    loc3 = new HRegionLocation(hri3, sn2);
    -109
    -110  // Replica stuff
    -111  private static final RegionInfo hri1r1 
    = RegionReplicaUtil.getRegionInfoForReplica(hri1, 1);
    -112  private static final RegionInfo hri1r2 
    = RegionReplicaUtil.getRegionInfoForReplica(hri1, 2);
    -113  private static final RegionInfo hri2r1 
    = RegionReplicaUtil.getRegionInfoForReplica(hri2, 1);
    -114  private static final RegionLocations 
    hrls1 = new RegionLocations(new HRegionLocation(hri1, sn),
    -115  new HRegionLocation(hri1r1, sn2), 
    new HRegionLocation(hri1r2, sn3));
    -116  private static final RegionLocations 
    hrls2 = new RegionLocations(new HRegionLocation(hri2, sn2),
    -117  new HRegionLocation(hri2r1, 
    sn3));
    -118  private static final RegionLocations 
    hrls3 =
    -119  new RegionLocations(new 
    HRegionLocation(hri3, sn3), null);
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html 
    b/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
    index b809814..dc68db2 100644
    --- a/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
    +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
    @@ -219,7 +219,7 @@ extends Segment
    -close,
     compare,
     compareRows,
     decScannerCount,
     dump,
     getCellLength,
     getCellsCount,
     getCellSet,
     getComparator,
     getFirstAfter,
     getMemStoreLAB,
     getMinSequenceId,
     getScanner,
     getScanner,
     getScanners,
     getTimeRangeTracker,
     headSet, heapSize,
     heapSizeChange,
     incScannerCount,
     incSize,
     internalAdd,
     isEmpty,
     isTagsPresent,
     iterator, keySize,
     last,
     maybeCloneWithAllocator,
     setCellSet,
     shouldSeek,
     tailSet,
     updateMetaInfo, updateMetaInfo
    +close,
     compare,
     compareRows,
     decScannerCount,
     dump,
     getCellLength,
     getCellsCount,
     getCellSet,
     getComparator,
     getFirstAfter,
     getMemStoreLAB,
     getMinSequenceId,
     getScanner,
     getScanner,
     getScanners,
     getTimeRangeTracker,
     headSet, heapSize,
     heapSizeChange,
     incScannerCount,
     incSize,
     internalAdd,
     isEmpty,
     isTagsPresent,
     iterator, keySize,
     last,
     maybeCloneWithAllocator,
     setCellSet,
     shouldSeek,
     tailSet,
     updateMetaInfo, updateMetaInfo
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
     
    b/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
    index ad6efee..bbe4de7 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
    @@ -248,7 +248,7 @@ extends Segment
    -close,
     compare,
     compareRows,
     decScannerCount,
     dump,
     getCellLength,
     getCellsCount,
     getCellSet,
     getComparator,
     getFirstAfter,
     getMemStoreLAB,
     getMinSequenceId,
     getScanner,
     getScanner,
     getScanners,
     getTimeRangeTracker,
     headSet, heapSize,
     heapSizeChange,
     incScannerCount,
     incSize,
     internalAdd,
     isEmpty,
     isTagsPresent,
     iterator, keySize,
     last,
     maybeCloneWithAllocator,
     setCellSet,
     shouldSeek,
     tailSet,
     updateMetaInfo, updateMetaInfo
    +close,
     compare,
     compareRows,
     decScannerCount,
     dump,
     getCellLength,
     getCellsCount,
     getCellSet,
     getComparator,
     getFirstAfter,
     getMemStoreLAB,
     getMinSequenceId,
     getScanner,
     getScanner,
     getScanners,
     getTimeRangeTracker,
     headSet, heapSize,
     heapSizeChange,
     incScannerCount,
     incSize,
     internalAdd,
     isEmpty,
     isTagsPresent,
     iterator, keySize,
     last,
     maybeCloneWithAllocator,
     setCellSet,
     shouldSeek,
     tailSet,
     updateMetaInfo, updateMetaInfo
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
     
    b/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
    index 0b339b8..a79894f 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
    @@ -18,7 +18,7 @@
     catch(err) {
     }
     //-->
    -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10};
    +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10};
     var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
    Methods"],8:["t4","Concrete Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
    @@ -120,7 +120,7 @@ var activeTableTab = "activeTableTab";
     
     
     @InterfaceAudience.Private
    -public class CellChunkImmutableSegment
    +public class CellChunkImmutableSegment
     extends ImmutableSegment
     CellChunkImmutableSegment extends the API supported by a Segment,
      and ImmutableSegment. This 
    immutable segment is working with CellSet with
    @@ -224,22 +224,26 @@ extends canBeFlattened()
     
     
    +private Cell
    +copyCellIntoMSLAB(Cellcell)
    +
    +
     private int
     createCellReference(ByteBufferKeyValuecell,
    http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
     title="class or interface in java.nio">ByteBufferidxBuffer,
    intidxOffset)
     
    -
    +
     protected long
     indexEntrySize()
     
    -
    +
     private void
     initializeCellSet(intnumOfCells,
      MemStoreSegmentsIteratoriterator,
      MemStoreCompactionStrategy.Actionaction)
     
    -
    +
     private void
     reinitializeCellSet(intnumOfCells,
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
     
    b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
    index 19ab2eb..468f000 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
    @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -static class BucketCache.BucketEntry
    +static class BucketCache.BucketEntry
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
     title="class or interface in java.io">Serializable
     Item in cache. We expect this to be where most memory goes. 
    Java uses 8
    @@ -288,7 +288,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     serialVersionUID
    -private static finallong serialVersionUID
    +private static finallong serialVersionUID
     
     See Also:
     Constant
     Field Values
    @@ -301,7 +301,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     COMPARATOR
    -static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/Comparator.html?is-external=true;
     title="class or interface in java.util">ComparatorBucketCache.BucketEntry COMPARATOR
    +static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/Comparator.html?is-external=true;
     title="class or interface in java.util">ComparatorBucketCache.BucketEntry COMPARATOR
     
     
     
    @@ -310,7 +310,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     offsetBase
    -privateint offsetBase
    +privateint offsetBase
     
     
     
    @@ -319,7 +319,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     length
    -privateint length
    +privateint length
     
     
     
    @@ -328,7 +328,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     offset1
    -privatebyte offset1
    +privatebyte offset1
     
     
     
    @@ -337,7 +337,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     deserialiserIndex
    -byte deserialiserIndex
    +byte deserialiserIndex
     
     
     
    @@ -346,7 +346,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     accessCounter
    -private volatilelong accessCounter
    +private volatilelong accessCounter
     
     
     
    @@ -355,7 +355,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     priority
    -privateBlockPriority priority
    +privateBlockPriority priority
     
     
     
    @@ -364,7 +364,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     markedForEvict
    -private volatileboolean markedForEvict
    +private volatileboolean markedForEvict
     
     
     
    @@ -373,7 +373,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     refCount
    -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true;
     title="class or interface in java.util.concurrent.atomic">AtomicInteger refCount
    +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true;
     title="class or interface in java.util.concurrent.atomic">AtomicInteger refCount
     
     
     
    @@ -382,7 +382,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     cachedTime
    -private finallong cachedTime
    +private finallong cachedTime
     Time this block was cached.  Presumes we are created just 
    before we are added to the cache.
     
     
    @@ -400,7 +400,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     BucketEntry
    -BucketEntry(longoffset,
    +BucketEntry(longoffset,
     intlength,
     longaccessCounter,
     booleaninMemory)
    @@ -420,7 +420,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     offset
    -longoffset()
    +longoffset()
     
     
     
    @@ -429,7 +429,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     setOffset
    -privatevoidsetOffset(longvalue)
    +privatevoidsetOffset(longvalue)
     
     
     
    @@ -438,7 +438,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     getLength
    -publicintgetLength()
    +publicintgetLength()
     
     
     
    @@ -447,7 +447,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
     
     
     deserializerReference
    -protectedCacheableDeserializerCacheabledeserializerReference(UniqueIndexMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
     title="class or interface in 
    java.lang">IntegerdeserialiserMap)
    +protectedCacheableDeserializerCacheabledeserializerReference(UniqueIndexMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
    index 3400507..2baa140 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
    @@ -28,3034 +28,2926 @@
     020import static 
    org.apache.hadoop.hbase.HConstants.EMPTY_BYTE_ARRAY;
     021import static 
    org.apache.hadoop.hbase.Tag.TAG_LENGTH_SIZE;
     022
    -023import 
    com.google.common.annotations.VisibleForTesting;
    -024
    -025import java.io.DataOutput;
    -026import java.io.DataOutputStream;
    -027import java.io.IOException;
    -028import java.io.OutputStream;
    -029import java.math.BigDecimal;
    -030import java.nio.ByteBuffer;
    -031import java.util.ArrayList;
    -032import java.util.Iterator;
    -033import java.util.List;
    -034import java.util.Optional;
    -035
    -036import 
    org.apache.hadoop.hbase.KeyValue.Type;
    -037import 
    org.apache.hadoop.hbase.filter.ByteArrayComparable;
    -038import 
    org.apache.hadoop.hbase.io.HeapSize;
    -039import 
    org.apache.hadoop.hbase.io.TagCompressionContext;
    -040import 
    org.apache.hadoop.hbase.io.util.Dictionary;
    -041import 
    org.apache.hadoop.hbase.io.util.StreamUtils;
    -042import 
    org.apache.hadoop.hbase.util.ByteBufferUtils;
    -043import 
    org.apache.hadoop.hbase.util.ByteRange;
    -044import 
    org.apache.hadoop.hbase.util.Bytes;
    -045import 
    org.apache.hadoop.hbase.util.ClassSize;
    -046import 
    org.apache.yetus.audience.InterfaceAudience;
    -047
    -048
    -049/**
    -050 * Utility methods helpful slinging 
    {@link Cell} instances. It has more powerful and
    -051 * rich set of APIs than those in {@link 
    CellUtil} for internal usage.
    -052 */
    -053@InterfaceAudience.Private
    -054public final class PrivateCellUtil {
    -055
    -056  /**
    -057   * Private constructor to keep this 
    class from being instantiated.
    -058   */
    -059  private PrivateCellUtil() {
    -060  }
    +023import java.io.DataOutput;
    +024import java.io.DataOutputStream;
    +025import java.io.IOException;
    +026import java.io.OutputStream;
    +027import java.math.BigDecimal;
    +028import java.nio.ByteBuffer;
    +029import java.util.ArrayList;
    +030import java.util.Iterator;
    +031import java.util.List;
    +032import java.util.Optional;
    +033import 
    org.apache.hadoop.hbase.KeyValue.Type;
    +034import 
    org.apache.hadoop.hbase.filter.ByteArrayComparable;
    +035import 
    org.apache.hadoop.hbase.io.HeapSize;
    +036import 
    org.apache.hadoop.hbase.io.TagCompressionContext;
    +037import 
    org.apache.hadoop.hbase.io.util.Dictionary;
    +038import 
    org.apache.hadoop.hbase.io.util.StreamUtils;
    +039import 
    org.apache.hadoop.hbase.util.ByteBufferUtils;
    +040import 
    org.apache.hadoop.hbase.util.ByteRange;
    +041import 
    org.apache.hadoop.hbase.util.Bytes;
    +042import 
    org.apache.hadoop.hbase.util.ClassSize;
    +043import 
    org.apache.yetus.audience.InterfaceAudience;
    +044
    +045import 
    org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
    +046
    +047/**
    +048 * Utility methods helpful slinging 
    {@link Cell} instances. It has more powerful and
    +049 * rich set of APIs than those in {@link 
    CellUtil} for internal usage.
    +050 */
    +051@InterfaceAudience.Private
    +052public final class PrivateCellUtil {
    +053
    +054  /**
    +055   * Private constructor to keep this 
    class from being instantiated.
    +056   */
    +057  private PrivateCellUtil() {
    +058  }
    +059
    +060  /*** ByteRange 
    ***/
     061
    -062  /*** ByteRange 
    ***/
    -063
    -064  public static ByteRange 
    fillRowRange(Cell cell, ByteRange range) {
    -065return range.set(cell.getRowArray(), 
    cell.getRowOffset(), cell.getRowLength());
    -066  }
    -067
    -068  public static ByteRange 
    fillFamilyRange(Cell cell, ByteRange range) {
    -069return 
    range.set(cell.getFamilyArray(), cell.getFamilyOffset(), 
    cell.getFamilyLength());
    -070  }
    -071
    -072  public static ByteRange 
    fillQualifierRange(Cell cell, ByteRange range) {
    -073return 
    range.set(cell.getQualifierArray(), cell.getQualifierOffset(),
    -074  cell.getQualifierLength());
    -075  }
    -076
    -077  public static ByteRange 
    fillValueRange(Cell cell, ByteRange range) {
    -078return 
    range.set(cell.getValueArray(), cell.getValueOffset(), 
    cell.getValueLength());
    -079  }
    -080
    -081  public static ByteRange 
    fillTagRange(Cell cell, ByteRange range) {
    -082return range.set(cell.getTagsArray(), 
    cell.getTagsOffset(), cell.getTagsLength());
    -083  }
    +062  public static ByteRange 
    fillRowRange(Cell cell, ByteRange range) {
    +063return range.set(cell.getRowArray(), 
    cell.getRowOffset(), cell.getRowLength());
    +064  }
    +065
    +066  public static ByteRange 
    fillFamilyRange(Cell cell, ByteRange range) {
    +067return 
    range.set(cell.getFamilyArray(), 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
    --
    diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html 
    b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
    index 9d49b9a..c36fdce 100644
    --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
    +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
    @@ -165,3380 +165,3375 @@
     157import 
    org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy;
     158import 
    org.apache.hadoop.hbase.regionserver.compactions.FIFOCompactionPolicy;
     159import 
    org.apache.hadoop.hbase.replication.ReplicationException;
    -160import 
    org.apache.hadoop.hbase.replication.ReplicationFactory;
    -161import 
    org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
    -162import 
    org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
    -163import 
    org.apache.hadoop.hbase.replication.ReplicationQueuesZKImpl;
    -164import 
    org.apache.hadoop.hbase.replication.master.ReplicationPeerConfigUpgrader;
    -165import 
    org.apache.hadoop.hbase.replication.regionserver.Replication;
    -166import 
    org.apache.hadoop.hbase.security.AccessDeniedException;
    -167import 
    org.apache.hadoop.hbase.security.UserProvider;
    -168import 
    org.apache.hadoop.hbase.trace.TraceUtil;
    -169import 
    org.apache.hadoop.hbase.util.Addressing;
    -170import 
    org.apache.hadoop.hbase.util.Bytes;
    -171import 
    org.apache.hadoop.hbase.util.CompressionTest;
    -172import 
    org.apache.hadoop.hbase.util.EncryptionTest;
    -173import 
    org.apache.hadoop.hbase.util.FSUtils;
    -174import 
    org.apache.hadoop.hbase.util.HFileArchiveUtil;
    -175import 
    org.apache.hadoop.hbase.util.HasThread;
    -176import 
    org.apache.hadoop.hbase.util.IdLock;
    -177import 
    org.apache.hadoop.hbase.util.ModifyRegionUtils;
    -178import 
    org.apache.hadoop.hbase.util.Pair;
    -179import 
    org.apache.hadoop.hbase.util.Threads;
    -180import 
    org.apache.hadoop.hbase.util.VersionInfo;
    -181import 
    org.apache.hadoop.hbase.util.ZKDataMigrator;
    -182import 
    org.apache.hadoop.hbase.zookeeper.LoadBalancerTracker;
    -183import 
    org.apache.hadoop.hbase.zookeeper.MasterAddressTracker;
    -184import 
    org.apache.hadoop.hbase.zookeeper.MasterMaintenanceModeTracker;
    -185import 
    org.apache.hadoop.hbase.zookeeper.RegionNormalizerTracker;
    -186import 
    org.apache.hadoop.hbase.zookeeper.ZKClusterId;
    -187import 
    org.apache.hadoop.hbase.zookeeper.ZKUtil;
    -188import 
    org.apache.hadoop.hbase.zookeeper.ZKWatcher;
    -189import 
    org.apache.hadoop.hbase.zookeeper.ZNodePaths;
    -190import 
    org.apache.yetus.audience.InterfaceAudience;
    -191import 
    org.apache.zookeeper.KeeperException;
    -192import org.eclipse.jetty.server.Server;
    -193import 
    org.eclipse.jetty.server.ServerConnector;
    -194import 
    org.eclipse.jetty.servlet.ServletHolder;
    -195import 
    org.eclipse.jetty.webapp.WebAppContext;
    -196import org.slf4j.Logger;
    -197import org.slf4j.LoggerFactory;
    -198
    -199import 
    org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
    -200import 
    org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;
    -201import 
    org.apache.hadoop.hbase.shaded.com.google.common.collect.Maps;
    -202import 
    org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
    -203import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState;
    -204import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionServerInfo;
    -205import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas;
    -206import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceViolationPolicy;
    -207import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;
    -208import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
    -209
    -210/**
    -211 * HMaster is the "master server" for 
    HBase. An HBase cluster has one active
    -212 * master.  If many masters are started, 
    all compete.  Whichever wins goes on to
    -213 * run the cluster.  All others park 
    themselves in their constructor until
    -214 * master or cluster shutdown or until 
    the active master loses its lease in
    -215 * zookeeper.  Thereafter, all running 
    master jostle to take over master role.
    -216 *
    -217 * pThe Master can be asked 
    shutdown the cluster. See {@link #shutdown()}.  In
    -218 * this case it will tell all 
    regionservers to go down and then wait on them
    -219 * all reporting in that they are down.  
    This master will then shut itself down.
    +160import 
    org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
    +161import 
    org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
    +162import 
    org.apache.hadoop.hbase.replication.master.ReplicationPeerConfigUpgrader;
    +163import 
    org.apache.hadoop.hbase.replication.regionserver.Replication;
    +164import 
    org.apache.hadoop.hbase.security.AccessDeniedException;
    +165import 
    org.apache.hadoop.hbase.security.UserProvider;
    +166import 
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
    index 6fecbc9..2accda0 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
    @@ -34,4140 +34,4141 @@
     026import 
    java.nio.charset.StandardCharsets;
     027import java.util.ArrayList;
     028import java.util.Arrays;
    -029import java.util.Collection;
    -030import java.util.EnumSet;
    -031import java.util.HashMap;
    -032import java.util.Iterator;
    -033import java.util.LinkedList;
    -034import java.util.List;
    -035import java.util.Map;
    -036import java.util.Set;
    -037import java.util.concurrent.Callable;
    -038import 
    java.util.concurrent.ExecutionException;
    -039import java.util.concurrent.Future;
    -040import java.util.concurrent.TimeUnit;
    -041import 
    java.util.concurrent.TimeoutException;
    -042import 
    java.util.concurrent.atomic.AtomicInteger;
    -043import 
    java.util.concurrent.atomic.AtomicReference;
    -044import java.util.regex.Pattern;
    -045import java.util.stream.Collectors;
    -046import java.util.stream.Stream;
    -047import 
    org.apache.hadoop.conf.Configuration;
    -048import 
    org.apache.hadoop.hbase.Abortable;
    -049import 
    org.apache.hadoop.hbase.CacheEvictionStats;
    -050import 
    org.apache.hadoop.hbase.CacheEvictionStatsBuilder;
    -051import 
    org.apache.hadoop.hbase.ClusterMetrics.Option;
    -052import 
    org.apache.hadoop.hbase.ClusterStatus;
    -053import 
    org.apache.hadoop.hbase.DoNotRetryIOException;
    -054import 
    org.apache.hadoop.hbase.HBaseConfiguration;
    -055import 
    org.apache.hadoop.hbase.HConstants;
    -056import 
    org.apache.hadoop.hbase.HRegionInfo;
    -057import 
    org.apache.hadoop.hbase.HRegionLocation;
    -058import 
    org.apache.hadoop.hbase.HTableDescriptor;
    -059import 
    org.apache.hadoop.hbase.MasterNotRunningException;
    -060import 
    org.apache.hadoop.hbase.MetaTableAccessor;
    -061import 
    org.apache.hadoop.hbase.NamespaceDescriptor;
    -062import 
    org.apache.hadoop.hbase.NamespaceNotFoundException;
    -063import 
    org.apache.hadoop.hbase.NotServingRegionException;
    -064import 
    org.apache.hadoop.hbase.RegionLoad;
    -065import 
    org.apache.hadoop.hbase.RegionLocations;
    -066import 
    org.apache.hadoop.hbase.ServerName;
    -067import 
    org.apache.hadoop.hbase.TableExistsException;
    -068import 
    org.apache.hadoop.hbase.TableName;
    -069import 
    org.apache.hadoop.hbase.TableNotDisabledException;
    -070import 
    org.apache.hadoop.hbase.TableNotFoundException;
    -071import 
    org.apache.hadoop.hbase.UnknownRegionException;
    -072import 
    org.apache.hadoop.hbase.ZooKeeperConnectionException;
    -073import 
    org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil;
    -074import 
    org.apache.hadoop.hbase.client.replication.TableCFs;
    -075import 
    org.apache.hadoop.hbase.client.security.SecurityCapability;
    -076import 
    org.apache.hadoop.hbase.exceptions.TimeoutIOException;
    -077import 
    org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
    -078import 
    org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
    -079import 
    org.apache.hadoop.hbase.ipc.HBaseRpcController;
    -080import 
    org.apache.hadoop.hbase.ipc.RpcControllerFactory;
    -081import 
    org.apache.hadoop.hbase.quotas.QuotaFilter;
    -082import 
    org.apache.hadoop.hbase.quotas.QuotaRetriever;
    -083import 
    org.apache.hadoop.hbase.quotas.QuotaSettings;
    -084import 
    org.apache.hadoop.hbase.regionserver.wal.FailedLogCloseException;
    -085import 
    org.apache.hadoop.hbase.replication.ReplicationException;
    -086import 
    org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
    -087import 
    org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
    -088import 
    org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
    -089import 
    org.apache.hadoop.hbase.snapshot.HBaseSnapshotException;
    -090import 
    org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
    -091import 
    org.apache.hadoop.hbase.snapshot.SnapshotCreationException;
    -092import 
    org.apache.hadoop.hbase.snapshot.UnknownSnapshotException;
    -093import 
    org.apache.hadoop.hbase.util.Addressing;
    -094import 
    org.apache.hadoop.hbase.util.Bytes;
    -095import 
    org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
    -096import 
    org.apache.hadoop.hbase.util.ForeignExceptionUtil;
    -097import 
    org.apache.hadoop.hbase.util.Pair;
    -098import 
    org.apache.hadoop.ipc.RemoteException;
    -099import 
    org.apache.hadoop.util.StringUtils;
    -100import 
    org.apache.yetus.audience.InterfaceAudience;
    -101import 
    org.apache.yetus.audience.InterfaceStability;
    -102import org.slf4j.Logger;
    -103import org.slf4j.LoggerFactory;
    -104
    -105import 
    org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
    -106import 
    org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html 
    b/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html
    index f60f387..2d6cdec 100644
    --- a/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html
    +++ b/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html
    @@ -18,8 +18,8 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":42,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":42,"i20":10,"i21":10,"i22":42,"i23":10,"i24":10,"i25":9,"i26":10};
    -var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
    Methods"],32:["t6","Deprecated Methods"]};
    +var methods = 
    {"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":42,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":42,"i28":42};
    +var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
    Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
     var tableTab = "tableTab";
    @@ -49,8 +49,8 @@ var activeTableTab = "activeTableTab";
     
     
     
    -PrevClass
    -NextClass
    +PrevClass
    +NextClass
     
     
     Frames
    @@ -74,7 +74,7 @@ var activeTableTab = "activeTableTab";
     
     
     Summary:
    -Nested|
    +Nested|
     Field|
     Constr|
     Method
    @@ -107,11 +107,21 @@ var activeTableTab = "activeTableTab";
     
     
     
    +
    +All Implemented Interfaces:
    +ClusterMetrics
    +
     
    +Deprecated.
    +As of release 2.0.0, this 
    will be removed in HBase 3.0.0
    + Use ClusterMetrics 
    instead.
    +
     
     @InterfaceAudience.Public
    -public class ClusterStatus
    -extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
    + http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true;
     title="class or interface in java.lang">@Deprecated
    +public class ClusterStatus
    +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
    +implements ClusterMetrics
     Status information on the HBase cluster.
      
      ClusterStatus provides clients with information such as:
    @@ -128,7 +138,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
      Regions in transition at master
      The unique cluster ID
      
    - ClusterStatus.Option provides a 
    way to get desired ClusterStatus information.
    + ClusterMetrics.Option provides a 
    way to get desired ClusterStatus information.
      The following codes will get all the cluster information.
      
      
    @@ -159,25 +169,13 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     
     
     Nested Class Summary
    -
    -Nested Classes
    -
    -Modifier and Type
    -Class and Description
    -
    -
    -static class
    -ClusterStatus.Builder
    -Builder for construct a ClusterStatus.
    -
    -
    -
    -static class
    -ClusterStatus.Option
    -Kinds of ClusterStatus
    -
    -
    -
    +
    +
    +
    +
    +Nested classes/interfaces inherited from 
    interfaceorg.apache.hadoop.hbase.ClusterMetrics
    +ClusterMetrics.Option
    +
     
     
     
    @@ -193,48 +191,16 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
     Field and Description
     
     
    -private http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
     title="class or interface in java.util">ListServerName
    -backupMasters
    +private ClusterMetrics
    +metrics
    +Deprecated.
    +
     
     
    -private http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
     title="class or interface in java.lang">Boolean
    -balancerOn
    -
    -
    -private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
    -clusterId
    -
    -
    -private http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
     title="class or interface in java.util">ListServerName
    -deadServers
    -
    -
    -private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String
    -hbaseVersion
    -
    -
    -private http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
     title="class or interface in java.util">ListRegionState
    -intransition
    -
    -
    -private http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
     title="class or interface in java.util">MapServerName,ServerLoad
    -liveServers
    -
    -
    -private ServerName
    -master
    -
    -
    -private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String[]
    -masterCoprocessors
    -
    -
    -private int
    -masterInfoPort
    -
    -
     private static byte
    -VERSION
    +VERSION
    +Deprecated.
    +
     
     
     
    @@ -248,12 +214,15 @@ extends 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html 
    b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
    index 129390f..a21c2b6 100644
    --- a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
    +++ b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
    @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
     
     
     @InterfaceAudience.Private
    -public final class BackupSystemTable
    +public final class BackupSystemTable
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Object
     implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true;
     title="class or interface in java.io">Closeable
     This class provides API to access backup system table
    @@ -234,7 +234,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     INCR_BACKUP_SET
     
     
    -private static 
    org.apache.commons.logging.Log
    +private static org.slf4j.Logger
     LOG
     
     
    @@ -953,7 +953,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     LOG
    -private static finalorg.apache.commons.logging.Log LOG
    +private static finalorg.slf4j.Logger LOG
     
     
     
    @@ -962,7 +962,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     tableName
    -privateTableName tableName
    +privateTableName tableName
     
     
     
    @@ -971,7 +971,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     SESSIONS_FAMILY
    -static finalbyte[] SESSIONS_FAMILY
    +static finalbyte[] SESSIONS_FAMILY
     Stores backup sessions (contexts)
     
     
    @@ -981,7 +981,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     META_FAMILY
    -static finalbyte[] META_FAMILY
    +static finalbyte[] META_FAMILY
     Stores other meta
     
     
    @@ -991,7 +991,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     BULK_LOAD_FAMILY
    -static finalbyte[] BULK_LOAD_FAMILY
    +static finalbyte[] BULK_LOAD_FAMILY
     
     
     
    @@ -1000,7 +1000,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     connection
    -private finalConnection connection
    +private finalConnection connection
     Connection to HBase cluster, shared among all 
    instances
     
     
    @@ -1010,7 +1010,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     BACKUP_INFO_PREFIX
    -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String BACKUP_INFO_PREFIX
    +private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String BACKUP_INFO_PREFIX
     
     See Also:
     Constant
     Field Values
    @@ -1023,7 +1023,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     START_CODE_ROW
    -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String START_CODE_ROW
    +private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String START_CODE_ROW
     
     See Also:
     Constant
     Field Values
    @@ -1036,7 +1036,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     ACTIVE_SESSION_ROW
    -private static finalbyte[] ACTIVE_SESSION_ROW
    +private static finalbyte[] ACTIVE_SESSION_ROW
     
     
     
    @@ -1045,7 +1045,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     ACTIVE_SESSION_COL
    -private static finalbyte[] ACTIVE_SESSION_COL
    +private static finalbyte[] ACTIVE_SESSION_COL
     
     
     
    @@ -1054,7 +1054,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     ACTIVE_SESSION_YES
    -private static finalbyte[] ACTIVE_SESSION_YES
    +private static finalbyte[] ACTIVE_SESSION_YES
     
     
     
    @@ -1063,7 +1063,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     ACTIVE_SESSION_NO
    -private static finalbyte[] ACTIVE_SESSION_NO
    +private static finalbyte[] ACTIVE_SESSION_NO
     
     
     
    @@ -1072,7 +1072,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     INCR_BACKUP_SET
    -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String INCR_BACKUP_SET
    +private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
     title="class or interface in java.lang">String INCR_BACKUP_SET
     
     See Also:
     Constant
     Field Values
    @@ -1085,7 +1085,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
     
     
     TABLE_RS_LOG_MAP_PREFIX
    -private static 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dad9a249/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
     
    b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
    index 067d702..d42cfb1 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
    @@ -137,7 +137,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -public static class ThriftServerRunner.IOErrorWithCause
    +public static class ThriftServerRunner.IOErrorWithCause
     extends org.apache.hadoop.hbase.thrift.generated.IOError
     
     See Also:
    @@ -273,7 +273,7 @@ extends 
    org.apache.hadoop.hbase.thrift.generated.IOError
     
     
     cause
    -privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
     title="class or interface in java.lang">Throwable cause
    +privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
     title="class or interface in java.lang">Throwable cause
     
     
     
    @@ -290,7 +290,7 @@ extends 
    org.apache.hadoop.hbase.thrift.generated.IOError
     
     
     IOErrorWithCause
    -publicIOErrorWithCause(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
     title="class or interface in java.lang">Throwablecause)
    +publicIOErrorWithCause(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
     title="class or interface in java.lang">Throwablecause)
     
     
     
    @@ -307,7 +307,7 @@ extends 
    org.apache.hadoop.hbase.thrift.generated.IOError
     
     
     getCause
    -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
     title="class or interface in java.lang">ThrowablegetCause()
    +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
     title="class or interface in java.lang">ThrowablegetCause()
     
     Overrides:
     http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getCause--;
     title="class or interface in java.lang">getCausein 
    classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
     title="class or interface in java.lang">Throwable
    @@ -320,7 +320,7 @@ extends 
    org.apache.hadoop.hbase.thrift.generated.IOError
     
     
     equals
    -publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Objectother)
    +publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
     title="class or interface in java.lang">Objectother)
     
     Overrides:
     equalsin 
    classorg.apache.hadoop.hbase.thrift.generated.IOError
    @@ -333,7 +333,7 @@ extends 
    org.apache.hadoop.hbase.thrift.generated.IOError
     
     
     hashCode
    -publicinthashCode()
    +publicinthashCode()
     
     Overrides:
     hashCodein 
    classorg.apache.hadoop.hbase.thrift.generated.IOError
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dad9a249/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html 
    b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
    index cd86385..b8aeaa7 100644
    --- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
    +++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
    @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -static enum ThriftServerRunner.ImplType
    +static enum ThriftServerRunner.ImplType
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
     title="class or interface in java.lang">EnumThriftServerRunner.ImplType
     An enum of server implementation selections
     
    @@ -281,7 +281,7 @@ the order they are declared.
     
     
     HS_HA
    -public static finalThriftServerRunner.ImplType 
    HS_HA
    +public static finalThriftServerRunner.ImplType 
    HS_HA
     
     
     
    @@ -290,7 +290,7 @@ the order they are declared.
     
     
     NONBLOCKING
    -public static finalThriftServerRunner.ImplType 
    NONBLOCKING
    +public static finalThriftServerRunner.ImplType 
    NONBLOCKING
     
     
     
    @@ -299,7 +299,7 @@ the order they are declared.
     
     
     THREAD_POOL
    -public static finalThriftServerRunner.ImplType 
    THREAD_POOL
    +public static finalThriftServerRunner.ImplType 
    THREAD_POOL
     
     
     
    @@ -308,7 +308,7 @@ the order they are declared.
     
     
     THREADED_SELECTOR
    -public static finalThriftServerRunner.ImplType 
    THREADED_SELECTOR
    +public static finalThriftServerRunner.ImplType 
    THREADED_SELECTOR
     
     
     
    @@ -325,7 +325,7 @@ the order they are declared.
     
     
     DEFAULT
    -public static finalThriftServerRunner.ImplType 
    DEFAULT
    +public static finalThriftServerRunner.ImplType 
    DEFAULT
     
     
     
    @@ -334,7 +334,7 @@ the order 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html
    --
    diff --git 
    a/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html 
    b/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html
    index 6646ba1..70481ce 100644
    --- 
    a/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html
    +++ 
    b/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html
    @@ -43,1651 +43,1652 @@
     035import java.io.OutputStream;
     036import java.io.Serializable;
     037import 
    java.io.UnsupportedEncodingException;
    -038import java.util.zip.GZIPInputStream;
    -039import java.util.zip.GZIPOutputStream;
    -040
    -041import org.apache.commons.logging.Log;
    -042import 
    org.apache.commons.logging.LogFactory;
    -043import 
    org.apache.yetus.audience.InterfaceAudience;
    -044
    -045/**
    -046 * Encodes and decodes to and from Base64 
    notation.
    -047 *
    -048 * p
    -049 * Homepage: a 
    href="http://iharder.net/base64"http://iharder.net/base64/a;.
    -050 * /p
    -051 *
    -052 * p
    -053 * Change Log:
    -054 * /p
    -055 * ul
    -056 *   liv2.2.1 - Fixed bug using 
    URL_SAFE and ORDERED encodings. Fixed bug
    -057 * when using very small files 
    (~lt; 40 bytes)./li
    -058 *   liv2.2 - Added some helper 
    methods for encoding/decoding directly from
    -059 * one file to the next. Also added a 
    main() method to support command
    -060 * line encoding/decoding from one 
    file to the next. Also added these
    -061 * Base64 dialects:
    -062 * ol
    -063 *   liThe default is RFC3548 
    format./li
    -064 *   liUsing Base64.URLSAFE 
    generates URL and file name friendly format as
    -065 * described in Section 4 of 
    RFC3548.
    -066 * 
    http://www.faqs.org/rfcs/rfc3548.html/li;
    -067 *   liUsing Base64.ORDERED 
    generates URL and file name friendly format
    -068 * that preserves lexical 
    ordering as described in
    -069 * 
    http://www.faqs.org/qa/rfcc-1940.html/li;
    -070 * /ol
    -071 * p
    -072 * Special thanks to Jim Kellerman at 
    a href="http://www.powerset.com/";
    -073 * http://www.powerset.com//a; 
    for contributing the new Base64 dialects.
    -074 *   /li
    -075 *
    -076 *   liv2.1 - Cleaned up javadoc 
    comments and unused variables and methods.
    -077 * Added some convenience methods for 
    reading and writing to and from files.
    -078 *   /li
    -079 *   liv2.0.2 - Now specifies 
    UTF-8 encoding in places where the code fails on
    -080 * systems with other encodings (like 
    EBCDIC)./li
    -081 *   liv2.0.1 - Fixed an error 
    when decoding a single byte, that is, when the
    -082 * encoded data was a single 
    byte./li
    -083 *   liv2.0 - I got rid of 
    methods that used booleans to set options. Now
    -084 * everything is more consolidated 
    and cleaner. The code now detects when
    -085 * data that's being decoded is 
    gzip-compressed and will decompress it
    -086 * automatically. Generally things 
    are cleaner. You'll probably have to
    -087 * change some method calls that you 
    were making to support the new options
    -088 * format (ttint/tts 
    that you "OR" together)./li
    -089 *   liv1.5.1 - Fixed bug when 
    decompressing and decoding to a byte[] using
    -090 * ttdecode( String s, 
    boolean gzipCompressed )/tt. Added the ability to
    -091 * "suspend" encoding in the Output 
    Stream so you can turn on and off the
    -092 * encoding if you need to embed 
    base64 data in an otherwise "normal" stream
    -093 * (like an XML file)./li
    -094 *   liv1.5 - Output stream pases 
    on flush() command but doesn't do anything
    -095 * itself. This helps when using GZIP 
    streams. Added the ability to
    -096 * GZip-compress objects before 
    encoding them./li
    -097 *   liv1.4 - Added helper 
    methods to read/write files./li
    -098 *   liv1.3.6 - Fixed 
    OutputStream.flush() so that 'position' is reset./li
    -099 *   liv1.3.5 - Added flag to 
    turn on and off line breaks. Fixed bug in input
    -100 * stream where last buffer being 
    read, if not completely full, was not
    -101 * returned./li
    -102 *   liv1.3.4 - Fixed when 
    "improperly padded stream" error was thrown at the
    -103 * wrong time./li
    -104 *   liv1.3.3 - Fixed I/O streams 
    which were totally messed up./li
    -105 * /ul
    -106 *
    -107 * p
    -108 * I am placing this code in the Public 
    Domain. Do with it as you will. This
    -109 * software comes with no guarantees or 
    warranties but with plenty of
    -110 * well-wishing instead!
    -111 * p
    -112 * Please visit a 
    href="http://iharder.net/base64"http://iharder.net/base64/a;
    -113 * periodically to check for updates or 
    to contribute improvements.
    -114 * p
    -115 * author: Robert Harder, 
    r...@iharder.net
    -116 * br
    -117 * version: 2.2.1
    -118 */
    -119@InterfaceAudience.Public
    -120public class Base64 {
    -121
    -122  /*  P U B L I C   F I E L D S 
     */
    -123
    -124  /** No options specified. Value is 
    zero. */
    -125  public final static int NO_OPTIONS = 
    0;
    -126
    -127  

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
    index 5e45072..34535d8 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
    @@ -50,754 +50,751 @@
     042import 
    org.apache.hadoop.hbase.CallQueueTooBigException;
     043import 
    org.apache.hadoop.hbase.CellScanner;
     044import 
    org.apache.hadoop.hbase.DoNotRetryIOException;
    -045import 
    org.apache.hadoop.hbase.HBaseInterfaceAudience;
    -046import 
    org.apache.hadoop.hbase.HConstants;
    -047import org.apache.hadoop.hbase.Server;
    -048import 
    org.apache.hadoop.hbase.conf.ConfigurationObserver;
    -049import 
    org.apache.hadoop.hbase.exceptions.RequestTooBigException;
    -050import 
    org.apache.hadoop.hbase.io.ByteBufferPool;
    -051import 
    org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
    -052import 
    org.apache.hadoop.hbase.monitoring.TaskMonitor;
    -053import 
    org.apache.hadoop.hbase.nio.ByteBuff;
    -054import 
    org.apache.hadoop.hbase.nio.MultiByteBuff;
    -055import 
    org.apache.hadoop.hbase.nio.SingleByteBuff;
    -056import 
    org.apache.hadoop.hbase.regionserver.RSRpcServices;
    -057import 
    org.apache.hadoop.hbase.security.SaslUtil;
    -058import 
    org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection;
    -059import 
    org.apache.hadoop.hbase.security.User;
    -060import 
    org.apache.hadoop.hbase.security.UserProvider;
    -061import 
    org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
    -062import 
    org.apache.hadoop.hbase.util.Pair;
    -063import 
    org.apache.hadoop.security.UserGroupInformation;
    -064import 
    org.apache.hadoop.security.authorize.AuthorizationException;
    -065import 
    org.apache.hadoop.security.authorize.PolicyProvider;
    -066import 
    org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
    -067import 
    org.apache.hadoop.security.token.SecretManager;
    -068import 
    org.apache.hadoop.security.token.TokenIdentifier;
    -069import 
    org.apache.yetus.audience.InterfaceAudience;
    -070import 
    org.apache.yetus.audience.InterfaceStability;
    -071
    -072import 
    org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
    -073import 
    org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
    -074import 
    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
    -075import 
    org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
    -076import 
    org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
    -077import 
    org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
    -078import 
    org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
    -079import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
    -080import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader;
    -081
    -082/**
    -083 * An RPC server that hosts protobuf 
    described Services.
    -084 *
    -085 */
    -086@InterfaceAudience.LimitedPrivate({HBaseInterfaceAudience.COPROC,
     HBaseInterfaceAudience.PHOENIX})
    -087@InterfaceStability.Evolving
    -088public abstract class RpcServer 
    implements RpcServerInterface,
    -089ConfigurationObserver {
    -090  // LOG is being used in CallRunner and 
    the log level is being changed in tests
    -091  public static final Log LOG = 
    LogFactory.getLog(RpcServer.class);
    -092  protected static final 
    CallQueueTooBigException CALL_QUEUE_TOO_BIG_EXCEPTION
    -093  = new CallQueueTooBigException();
    +045import 
    org.apache.hadoop.hbase.HConstants;
    +046import org.apache.hadoop.hbase.Server;
    +047import 
    org.apache.hadoop.hbase.conf.ConfigurationObserver;
    +048import 
    org.apache.hadoop.hbase.exceptions.RequestTooBigException;
    +049import 
    org.apache.hadoop.hbase.io.ByteBufferPool;
    +050import 
    org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
    +051import 
    org.apache.hadoop.hbase.monitoring.TaskMonitor;
    +052import 
    org.apache.hadoop.hbase.nio.ByteBuff;
    +053import 
    org.apache.hadoop.hbase.nio.MultiByteBuff;
    +054import 
    org.apache.hadoop.hbase.nio.SingleByteBuff;
    +055import 
    org.apache.hadoop.hbase.regionserver.RSRpcServices;
    +056import 
    org.apache.hadoop.hbase.security.SaslUtil;
    +057import 
    org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection;
    +058import 
    org.apache.hadoop.hbase.security.User;
    +059import 
    org.apache.hadoop.hbase.security.UserProvider;
    +060import 
    org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
    +061import 
    org.apache.hadoop.hbase.util.Pair;
    +062import 
    org.apache.hadoop.security.UserGroupInformation;
    +063import 
    org.apache.hadoop.security.authorize.AuthorizationException;
    +064import 
    org.apache.hadoop.security.authorize.PolicyProvider;
    +065import 
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
    index dd54dd2..a9c1142 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
    @@ -388,195 +388,196 @@
     380
     381  @Override
     382  public DatanodeInfo[] getPipeline() {
    -383return locations;
    -384  }
    -385
    -386  private void 
    flushBuffer(CompletableFutureLong future, ByteBuf dataBuf,
    -387  long nextPacketOffsetInBlock, 
    boolean syncBlock) {
    -388int dataLen = 
    dataBuf.readableBytes();
    -389int chunkLen = 
    summer.getBytesPerChecksum();
    -390int trailingPartialChunkLen = dataLen 
    % chunkLen;
    -391int numChecks = dataLen / chunkLen + 
    (trailingPartialChunkLen != 0 ? 1 : 0);
    -392int checksumLen = numChecks * 
    summer.getChecksumSize();
    -393ByteBuf checksumBuf = 
    alloc.directBuffer(checksumLen);
    -394
    summer.calculateChunkedSums(dataBuf.nioBuffer(), checksumBuf.nioBuffer(0, 
    checksumLen));
    -395
    checksumBuf.writerIndex(checksumLen);
    -396PacketHeader header = new 
    PacketHeader(4 + checksumLen + dataLen, nextPacketOffsetInBlock,
    -397nextPacketSeqno, false, dataLen, 
    syncBlock);
    -398int headerLen = 
    header.getSerializedSize();
    -399ByteBuf headerBuf = 
    alloc.buffer(headerLen);
    -400
    header.putInBuffer(headerBuf.nioBuffer(0, headerLen));
    -401headerBuf.writerIndex(headerLen);
    -402Callback c = new Callback(future, 
    nextPacketOffsetInBlock + dataLen, datanodeList);
    -403waitingAckQueue.addLast(c);
    -404// recheck again after we pushed the 
    callback to queue
    -405if (state != State.STREAMING 
     waitingAckQueue.peekFirst() == c) {
    -406  future.completeExceptionally(new 
    IOException("stream already broken"));
    -407  // it's the one we have just pushed 
    or just a no-op
    -408  waitingAckQueue.removeFirst();
    -409  return;
    -410}
    -411datanodeList.forEach(ch - {
    -412  
    ch.write(headerBuf.retainedDuplicate());
    -413  
    ch.write(checksumBuf.retainedDuplicate());
    -414  
    ch.writeAndFlush(dataBuf.retainedDuplicate());
    -415});
    -416checksumBuf.release();
    -417headerBuf.release();
    -418dataBuf.release();
    -419nextPacketSeqno++;
    -420  }
    -421
    -422  private void 
    flush0(CompletableFutureLong future, boolean syncBlock) {
    -423if (state != State.STREAMING) {
    -424  future.completeExceptionally(new 
    IOException("stream already broken"));
    -425  return;
    -426}
    -427int dataLen = buf.readableBytes();
    -428if (dataLen == 
    trailingPartialChunkLength) {
    -429  // no new data
    -430  long lengthAfterFlush = 
    nextPacketOffsetInBlock + dataLen;
    -431  Callback lastFlush = 
    waitingAckQueue.peekLast();
    -432  if (lastFlush != null) {
    -433Callback c = new Callback(future, 
    lengthAfterFlush, Collections.emptyList());
    -434waitingAckQueue.addLast(c);
    -435// recheck here if we have 
    already removed the previous callback from the queue
    -436if (waitingAckQueue.peekFirst() 
    == c) {
    -437  // all previous callbacks have 
    been removed
    -438  // notice that this does mean 
    we will always win here because the background thread may
    -439  // have already started to mark 
    the future here as completed in the completed or failed
    -440  // methods but haven't removed 
    it from the queue yet. That's also why the removeFirst
    -441  // call below may be a no-op.
    -442  if (state != State.STREAMING) 
    {
    -443
    future.completeExceptionally(new IOException("stream already broken"));
    -444  } else {
    -445
    future.complete(lengthAfterFlush);
    -446  }
    -447  // it's the one we have just 
    pushed or just a no-op
    -448  
    waitingAckQueue.removeFirst();
    -449}
    -450  } else {
    -451// we must have acked all the 
    data so the ackedBlockLength must be same with
    -452// lengthAfterFlush
    -453
    future.complete(lengthAfterFlush);
    -454  }
    -455  return;
    -456}
    -457
    -458if (encryptor != null) {
    -459  ByteBuf encryptBuf = 
    alloc.directBuffer(dataLen);
    -460  buf.readBytes(encryptBuf, 
    trailingPartialChunkLength);
    -461  int toEncryptLength = dataLen - 
    trailingPartialChunkLength;
    -462  try {
    -463
    encryptor.encrypt(buf.nioBuffer(trailingPartialChunkLength, toEncryptLength),
    -464  
    encryptBuf.nioBuffer(trailingPartialChunkLength, toEncryptLength));
    -465  } catch (IOException e) {
    -466encryptBuf.release();
    -467
    future.completeExceptionally(e);
    -468

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
    index ae9896e..cb2d47f 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
    @@ -41,25 +41,25 @@
     033import java.util.TreeMap;
     034import java.util.TreeSet;
     035
    -036import 
    org.apache.commons.lang3.StringUtils;
    -037import org.apache.commons.logging.Log;
    -038import 
    org.apache.commons.logging.LogFactory;
    -039import 
    org.apache.hadoop.conf.Configuration;
    -040import org.apache.hadoop.fs.Path;
    -041import org.apache.hadoop.hbase.Cell;
    -042import 
    org.apache.hadoop.hbase.CellUtil;
    -043import 
    org.apache.hadoop.hbase.HBaseConfiguration;
    -044import 
    org.apache.hadoop.hbase.HColumnDescriptor;
    -045import 
    org.apache.hadoop.hbase.HTableDescriptor;
    -046import 
    org.apache.hadoop.hbase.NamespaceDescriptor;
    -047import 
    org.apache.hadoop.hbase.ServerName;
    -048import 
    org.apache.hadoop.hbase.TableName;
    -049import 
    org.apache.hadoop.hbase.backup.BackupInfo;
    -050import 
    org.apache.hadoop.hbase.backup.BackupInfo.BackupState;
    -051import 
    org.apache.hadoop.hbase.backup.BackupRestoreConstants;
    -052import 
    org.apache.hadoop.hbase.backup.BackupType;
    -053import 
    org.apache.hadoop.hbase.backup.util.BackupUtils;
    -054import 
    org.apache.yetus.audience.InterfaceAudience;
    +036import 
    org.apache.commons.lang3.ArrayUtils;
    +037import 
    org.apache.commons.lang3.StringUtils;
    +038import org.apache.commons.logging.Log;
    +039import 
    org.apache.commons.logging.LogFactory;
    +040import 
    org.apache.hadoop.conf.Configuration;
    +041import org.apache.hadoop.fs.Path;
    +042import org.apache.hadoop.hbase.Cell;
    +043import 
    org.apache.hadoop.hbase.CellUtil;
    +044import 
    org.apache.hadoop.hbase.HBaseConfiguration;
    +045import 
    org.apache.hadoop.hbase.HColumnDescriptor;
    +046import 
    org.apache.hadoop.hbase.HTableDescriptor;
    +047import 
    org.apache.hadoop.hbase.NamespaceDescriptor;
    +048import 
    org.apache.hadoop.hbase.ServerName;
    +049import 
    org.apache.hadoop.hbase.TableName;
    +050import 
    org.apache.hadoop.hbase.backup.BackupInfo;
    +051import 
    org.apache.hadoop.hbase.backup.BackupInfo.BackupState;
    +052import 
    org.apache.hadoop.hbase.backup.BackupRestoreConstants;
    +053import 
    org.apache.hadoop.hbase.backup.BackupType;
    +054import 
    org.apache.hadoop.hbase.backup.util.BackupUtils;
     055import 
    org.apache.hadoop.hbase.client.Admin;
     056import 
    org.apache.hadoop.hbase.client.Connection;
     057import 
    org.apache.hadoop.hbase.client.Delete;
    @@ -70,12 +70,12 @@
     062import 
    org.apache.hadoop.hbase.client.Scan;
     063import 
    org.apache.hadoop.hbase.client.SnapshotDescription;
     064import 
    org.apache.hadoop.hbase.client.Table;
    -065import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos;
    -066import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
    -067import 
    org.apache.hadoop.hbase.util.ArrayUtils;
    -068import 
    org.apache.hadoop.hbase.util.Bytes;
    -069import 
    org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
    -070import 
    org.apache.hadoop.hbase.util.Pair;
    +065import 
    org.apache.hadoop.hbase.util.Bytes;
    +066import 
    org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
    +067import 
    org.apache.hadoop.hbase.util.Pair;
    +068import 
    org.apache.yetus.audience.InterfaceAudience;
    +069import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos;
    +070import 
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
     071
     072/**
     073 * This class provides API to access 
    backup system tablebr
    @@ -585,1396 +585,1397 @@
     577try (Table table = 
    connection.getTable(tableName)) {
     578  Put put = 
    createPutForStartBackupSession();
     579  // First try to put if row does not 
    exist
    -580  if 
    (!table.checkAndPut(ACTIVE_SESSION_ROW, SESSIONS_FAMILY, ACTIVE_SESSION_COL, 
    null, put)) {
    -581// Row exists, try to put if 
    value == ACTIVE_SESSION_NO
    -582if 
    (!table.checkAndPut(ACTIVE_SESSION_ROW, SESSIONS_FAMILY, ACTIVE_SESSION_COL,
    -583  ACTIVE_SESSION_NO, put)) {
    -584  throw new IOException("There is 
    an active backup exclusive operation");
    -585}
    -586  }
    -587}
    -588  }
    -589
    -590  private Put 
    createPutForStartBackupSession() {
    -591Put put = new 
    Put(ACTIVE_SESSION_ROW);
    -592put.addColumn(SESSIONS_FAMILY, 
    ACTIVE_SESSION_COL, ACTIVE_SESSION_YES);
    -593return put;
    -594  }
    -595
    -596  public void 
    finishBackupExclusiveOperation() throws IOException {
    -597LOG.debug("Finish backup exclusive 
    operation");
    -598
    -599try (Table table = 
    connection.getTable(tableName)) {
    -600  Put put = 
    createPutForStopBackupSession();
    -601  if 
    (!table.checkAndPut(ACTIVE_SESSION_ROW, SESSIONS_FAMILY, 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html 
    b/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html
    index 2ec4ab5..21b91cb 100644
    --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html
    +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html
    @@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -PrevClass
    +PrevClass
     NextClass
     
     
    @@ -250,7 +250,7 @@ extends 
     
     
    -PrevClass
    +PrevClass
     NextClass
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
     
    b/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
    index 8964468..d941a19 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
    @@ -134,7 +134,7 @@ extends 
     Connection
     createConnection(org.apache.hadoop.conf.Configurationconf)
    -Creates a cluster connection using the passed 
    configuration.
    +Creates a cluster connection using the passed 
    Configuration.
     
     
     
    @@ -146,7 +146,7 @@ extends 
     Connection
     getConnection()
    -Be careful RPC'ing from a Coprocessor context.
    +Returns the hosts' Connection to the Cluster.
     
     
     
    @@ -268,22 +268,33 @@ extends 
     
     getConnection
    -ConnectiongetConnection()
    -Be careful RPC'ing from a Coprocessor context.
    - RPC's will fail, stall, retry, and/or crawl because the remote side is not 
    online, is
    - struggling or it is on the other side of a network partition. Any use of 
    Connection from
    - inside a Coprocessor must be able to handle all such hiccups.
    +ConnectiongetConnection()
    +Returns the hosts' Connection to the Cluster. Do not 
    close! This is a shared connection
    + with the hosting server. Throws http://docs.oracle.com/javase/8/docs/api/java/lang/UnsupportedOperationException.html?is-external=true;
     title="class or interface in 
    java.lang">UnsupportedOperationException if you try to close
    + or abort it.
     
    - Using a Connection to get at a local resource -- say a Region that is on 
    the local
    - Server or using Admin Interface from a Coprocessor hosted on the Master -- 
    will result in a
    - short-circuit of the RPC framework to make a direct invocation avoiding RPC.
    -
    - Note: If you want to create Connection with your own Configuration and NOT 
    use the RegionServer
    - Connection (though its cache of locations will be warm, and its life-cycle is 
    not the concern
    - of the CP), see createConnection(Configuration).
    + For light-weight usage only. Heavy-duty usage will pull down
    + the hosting RegionServer responsiveness as well as that of other Coprocessors 
    making use of
    + this Connection. Use to create table on start or to do administrative 
    operations. Coprocessors
    + should create their own Connections if heavy usage to avoid impinging on 
    hosting Server
    + operation. To create a Connection or if a Coprocessor requires a region with 
    a particular
    + Configuration, use ConnectionFactory or
    + createConnection(Configuration)}.
    +
    + Be aware that operations that make use of this Connection are executed as 
    the RegionServer
    + User, the hbase super user that started this server process. Exercise caution 
    running
    + operations as this User (See createConnection(Configuration)}
     to run as other than
    + the RegionServer User).
    +
    + Be careful RPC'ing from a Coprocessor context. RPC's will fail, stall, 
    retry, and/or crawl
    + because the remote side is not online, is struggling or it is on the other 
    side of a network
    + partition. Any use of Connection from inside a Coprocessor must be able to 
    handle all such
    + hiccups.
     
     Returns:
     The host's Connection to the Cluster.
    +See Also:
    +createConnection(Configuration)
     
     
     
    @@ -293,18 +304,26 @@ extends 
     
     createConnection
    -ConnectioncreateConnection(org.apache.hadoop.conf.Configurationconf)
    +ConnectioncreateConnection(org.apache.hadoop.conf.Configurationconf)
      throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
     title="class or interface in java.io">IOException
    -Creates a cluster connection using the passed configuration.
    - Using this Connection to get at a local resource -- say a Region that is 
    on the local
    - Server or using Admin Interface from a Coprocessor hosted on the Master -- 
    will result in a
    - short-circuit of the RPC framework to make a direct invocation avoiding RPC.
    - 
    - Note: HBase will NOT cache/maintain this Connection. If Coprocessors need to 
    cache and reuse
    - this connection, it has to be done by Coprocessors. 

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
    --
    diff --git 
    a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
     
    b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
    index c0771ec..fa6927a 100644
    --- 
    a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
    +++ 
    b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase - Exemplar for hbase-client archetype  
    Dependency Information
     
    @@ -147,7 +147,7 @@
     https://www.apache.org/;>The Apache Software 
    Foundation.
     All rights reserved.  
     
    -  Last Published: 
    2017-12-05
    +  Last Published: 
    2017-12-06
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
    --
    diff --git 
    a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
     
    b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
    index fdd16aa..3c1da71 100644
    --- 
    a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
    +++ 
    b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase - Exemplar for hbase-client archetype  Project 
    Dependency Management
     
    @@ -775,18 +775,24 @@
     test-jar
     https://www.apache.org/licenses/LICENSE-2.0.txt;>Apache License, Version 
    2.0
     
    +org.apache.hbase
    +http://hbase.apache.org/hbase-build-configuration/hbase-zookeeper;>hbase-zookeeper
    +3.0.0-SNAPSHOT
    +test-jar
    +https://www.apache.org/licenses/LICENSE-2.0.txt;>Apache License, Version 
    2.0
    +
     org.bouncycastle
     http://www.bouncycastle.org/java.html;>bcprov-jdk16
     1.46
     jar
     http://www.bouncycastle.org/licence.html;>Bouncy Castle 
    Licence
    -
    +
     org.hamcrest
     https://github.com/hamcrest/JavaHamcrest/hamcrest-core;>hamcrest-core
     1.3
     jar
     http://www.opensource.org/licenses/bsd-license.php;>New BSD 
    License
    -
    +
     org.mockito
     http://mockito.org;>mockito-core
     2.1.0
    @@ -804,7 +810,7 @@
     https://www.apache.org/;>The Apache Software 
    Foundation.
     All rights reserved.  
     
    -  Last Published: 
    2017-12-05
    +  Last Published: 
    2017-12-06
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
    --
    diff --git 
    a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html 
    b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
    index 2e2bdd1..943db8b 100644
    --- a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
    +++ b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase - Exemplar for hbase-client archetype  
    About
     
    @@ -119,7 +119,7 @@
     https://www.apache.org/;>The Apache Software 
    Foundation.
     All rights reserved.  
     
    -  Last Published: 
    2017-12-05
    +  Last Published: 
    2017-12-06
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
    --
    diff --git 
    a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
     
    b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
    index 232443d..c65702d 100644
    --- 
    a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
    +++ 
    b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase - Exemplar for hbase-client archetype  CI 
    Management
     
    @@ -126,7 +126,7 @@
     https://www.apache.org/;>The Apache Software 
    Foundation.
     All rights reserved.  
     
    -  Last Published: 
    2017-12-05
    +  Last Published: 
    2017-12-06
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/issue-tracking.html
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
    --
    diff --git 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
     
    b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
    index 698163a..238f6d6 100644
    --- 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
    +++ 
    b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase - Exemplar for hbase-shaded-client archetype  
    Reactor Dependency Convergence
     
    @@ -838,7 +838,7 @@
     https://www.apache.org/;>The Apache Software 
    Foundation.
     All rights reserved.  
     
    -  Last Published: 
    2017-12-02
    +  Last Published: 
    2017-12-03
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
    --
    diff --git 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
     
    b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
    index f630509..d188ea3 100644
    --- 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
    +++ 
    b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase - Exemplar for hbase-shaded-client archetype  
    Dependency Information
     
    @@ -147,7 +147,7 @@
     https://www.apache.org/;>The Apache Software 
    Foundation.
     All rights reserved.  
     
    -  Last Published: 
    2017-12-02
    +  Last Published: 
    2017-12-03
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
    --
    diff --git 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
     
    b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
    index faa52ff..41779ab 100644
    --- 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
    +++ 
    b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase - Exemplar for hbase-shaded-client archetype  
    Project Dependency Management
     
    @@ -804,7 +804,7 @@
     https://www.apache.org/;>The Apache Software 
    Foundation.
     All rights reserved.  
     
    -  Last Published: 
    2017-12-02
    +  Last Published: 
    2017-12-03
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
    --
    diff --git 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
     
    b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
    index dc4de1d..e9a84e4 100644
    --- 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
    +++ 
    b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase - Exemplar for hbase-shaded-client archetype  
    About
     
    @@ -119,7 +119,7 @@
     https://www.apache.org/;>The Apache Software 
    Foundation.
     All rights reserved.  
     
    -  Last Published: 
    2017-12-02
    +  Last Published: 
    2017-12-03
     
     
     
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/integration.html
    --
    diff --git 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/integration.html
     
    b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/integration.html
    index c8d64af..2906ba6 100644
    --- 
    a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/integration.html
    +++ 
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
    index 25e368d..d0f781f 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
    @@ -25,798 +25,798 @@
     017 */
     018package 
    org.apache.hadoop.hbase.io.asyncfs;
     019
    -020import static 
    org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelOption.CONNECT_TIMEOUT_MILLIS;
    -021import static 
    org.apache.hadoop.hbase.shaded.io.netty.handler.timeout.IdleState.READER_IDLE;
    -022import static 
    org.apache.hadoop.fs.CreateFlag.CREATE;
    -023import static 
    org.apache.hadoop.fs.CreateFlag.OVERWRITE;
    -024import static 
    org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createEncryptor;
    -025import static 
    org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.trySaslNegotiate;
    +020import static 
    org.apache.hadoop.fs.CreateFlag.CREATE;
    +021import static 
    org.apache.hadoop.fs.CreateFlag.OVERWRITE;
    +022import static 
    org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createEncryptor;
    +023import static 
    org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.trySaslNegotiate;
    +024import static 
    org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelOption.CONNECT_TIMEOUT_MILLIS;
    +025import static 
    org.apache.hadoop.hbase.shaded.io.netty.handler.timeout.IdleState.READER_IDLE;
     026import static 
    org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_SOCKET_TIMEOUT_KEY;
     027import static 
    org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_USE_DN_HOSTNAME;
     028import static 
    org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_USE_DN_HOSTNAME_DEFAULT;
     029import static 
    org.apache.hadoop.hdfs.protocol.datatransfer.BlockConstructionStage.PIPELINE_SETUP_CREATE;
     030
    -031import 
    org.apache.hadoop.hbase.shaded.com.google.common.base.Throwables;
    -032import 
    org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap;
    -033import 
    com.google.protobuf.CodedOutputStream;
    -034
    -035import 
    org.apache.hadoop.hbase.shaded.io.netty.bootstrap.Bootstrap;
    -036import 
    org.apache.hadoop.hbase.shaded.io.netty.buffer.ByteBuf;
    -037import 
    org.apache.hadoop.hbase.shaded.io.netty.buffer.ByteBufAllocator;
    -038import 
    org.apache.hadoop.hbase.shaded.io.netty.buffer.ByteBufOutputStream;
    -039import 
    org.apache.hadoop.hbase.shaded.io.netty.buffer.PooledByteBufAllocator;
    -040import 
    org.apache.hadoop.hbase.shaded.io.netty.channel.Channel;
    -041import 
    org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelFuture;
    -042import 
    org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelFutureListener;
    -043import 
    org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandler;
    -044import 
    org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext;
    -045import 
    org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelInitializer;
    -046import 
    org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelPipeline;
    -047import 
    org.apache.hadoop.hbase.shaded.io.netty.channel.EventLoop;
    -048import 
    org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHandler;
    -049import 
    org.apache.hadoop.hbase.shaded.io.netty.handler.codec.protobuf.ProtobufDecoder;
    -050import 
    org.apache.hadoop.hbase.shaded.io.netty.handler.codec.protobuf.ProtobufVarint32FrameDecoder;
    -051import 
    org.apache.hadoop.hbase.shaded.io.netty.handler.timeout.IdleStateEvent;
    -052import 
    org.apache.hadoop.hbase.shaded.io.netty.handler.timeout.IdleStateHandler;
    -053import 
    org.apache.hadoop.hbase.shaded.io.netty.util.concurrent.Future;
    -054import 
    org.apache.hadoop.hbase.shaded.io.netty.util.concurrent.FutureListener;
    -055import 
    org.apache.hadoop.hbase.shaded.io.netty.util.concurrent.Promise;
    -056
    -057import java.io.IOException;
    -058import 
    java.lang.reflect.InvocationTargetException;
    -059import java.lang.reflect.Method;
    -060import java.util.ArrayList;
    -061import java.util.EnumSet;
    -062import java.util.List;
    -063import java.util.concurrent.TimeUnit;
    -064
    -065import org.apache.commons.logging.Log;
    -066import 
    org.apache.commons.logging.LogFactory;
    -067import 
    org.apache.hadoop.conf.Configuration;
    -068import 
    org.apache.hadoop.crypto.CryptoProtocolVersion;
    -069import 
    org.apache.hadoop.crypto.Encryptor;
    -070import org.apache.hadoop.fs.CreateFlag;
    -071import org.apache.hadoop.fs.FileSystem;
    -072import 
    org.apache.hadoop.fs.FileSystemLinkResolver;
    -073import org.apache.hadoop.fs.Path;
    -074import 
    org.apache.hadoop.fs.UnresolvedLinkException;
    -075import 
    org.apache.hadoop.fs.permission.FsPermission;
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
    index d438f22..7c59e27 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
    @@ -1290,8 +1290,8 @@
     1282   CompactType 
    compactType) throws IOException {
     1283switch (compactType) {
     1284  case MOB:
    -1285
    compact(this.connection.getAdminForMaster(), getMobRegionInfo(tableName), 
    major,
    -1286  columnFamily);
    +1285
    compact(this.connection.getAdminForMaster(), 
    RegionInfo.createMobRegionInfo(tableName),
    +1286major, columnFamily);
     1287break;
     1288  case NORMAL:
     1289checkTableExists(tableName);
    @@ -3248,7 +3248,7 @@
     3240  new 
    CallableAdminProtos.GetRegionInfoResponse.CompactionState() {
     3241@Override
     3242public 
    AdminProtos.GetRegionInfoResponse.CompactionState call() throws Exception {
    -3243  RegionInfo info = 
    getMobRegionInfo(tableName);
    +3243  RegionInfo info = 
    RegionInfo.createMobRegionInfo(tableName);
     3244  GetRegionInfoRequest 
    request =
     3245
    RequestConverter.buildGetRegionInfoRequest(info.getRegionName(), true);
     3246  GetRegionInfoResponse 
    response = masterAdmin.getRegionInfo(rpcController, request);
    @@ -3312,7 +3312,7 @@
     3304}
     3305break;
     3306  default:
    -3307throw new 
    IllegalArgumentException("Unknowne compactType: " + compactType);
    +3307throw new 
    IllegalArgumentException("Unknown compactType: " + compactType);
     3308}
     3309if (state != null) {
     3310  return 
    ProtobufUtil.createCompactionState(state);
    @@ -3847,325 +3847,320 @@
     3839});
     3840  }
     3841
    -3842  private RegionInfo 
    getMobRegionInfo(TableName tableName) {
    -3843return 
    RegionInfoBuilder.newBuilder(tableName).setStartKey(Bytes.toBytes(".mob")).setRegionId(0)
    -3844.build();
    -3845  }
    -3846
    -3847  private RpcControllerFactory 
    getRpcControllerFactory() {
    -3848return this.rpcControllerFactory;
    -3849  }
    -3850
    -3851  @Override
    -3852  public void addReplicationPeer(String 
    peerId, ReplicationPeerConfig peerConfig, boolean enabled)
    -3853  throws IOException {
    -3854executeCallable(new 
    MasterCallableVoid(getConnection(), getRpcControllerFactory()) {
    -3855  @Override
    -3856  protected Void rpcCall() throws 
    Exception {
    -3857
    master.addReplicationPeer(getRpcController(),
    -3858  
    RequestConverter.buildAddReplicationPeerRequest(peerId, peerConfig, 
    enabled));
    -3859return null;
    -3860  }
    -3861});
    -3862  }
    -3863
    -3864  @Override
    -3865  public void 
    removeReplicationPeer(String peerId) throws IOException {
    -3866executeCallable(new 
    MasterCallableVoid(getConnection(), getRpcControllerFactory()) {
    -3867  @Override
    -3868  protected Void rpcCall() throws 
    Exception {
    -3869
    master.removeReplicationPeer(getRpcController(),
    -3870  
    RequestConverter.buildRemoveReplicationPeerRequest(peerId));
    -3871return null;
    -3872  }
    -3873});
    -3874  }
    -3875
    -3876  @Override
    -3877  public void 
    enableReplicationPeer(final String peerId) throws IOException {
    -3878executeCallable(new 
    MasterCallableVoid(getConnection(), getRpcControllerFactory()) {
    -3879  @Override
    -3880  protected Void rpcCall() throws 
    Exception {
    -3881
    master.enableReplicationPeer(getRpcController(),
    -3882  
    RequestConverter.buildEnableReplicationPeerRequest(peerId));
    -3883return null;
    -3884  }
    -3885});
    -3886  }
    -3887
    -3888  @Override
    -3889  public void 
    disableReplicationPeer(final String peerId) throws IOException {
    -3890executeCallable(new 
    MasterCallableVoid(getConnection(), getRpcControllerFactory()) {
    -3891  @Override
    -3892  protected Void rpcCall() throws 
    Exception {
    -3893
    master.disableReplicationPeer(getRpcController(),
    -3894  
    RequestConverter.buildDisableReplicationPeerRequest(peerId));
    -3895return null;
    -3896  }
    -3897});
    -3898  }
    -3899
    -3900  @Override
    -3901  public ReplicationPeerConfig 
    getReplicationPeerConfig(final String peerId) throws IOException {
    -3902return executeCallable(new 
    MasterCallableReplicationPeerConfig(getConnection(),
    -3903getRpcControllerFactory()) {
    -3904  @Override
    -3905  protected ReplicationPeerConfig 
    rpcCall() throws Exception {
    -3906GetReplicationPeerConfigResponse 
    response = master.getReplicationPeerConfig(
    -3907  getRpcController(), 
    

    [07/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9722a17/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
    --
    diff --git 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
     
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
    index 29ea7b3..6ed75c9 100644
    --- 
    a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
    +++ 
    b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
    @@ -1313,7093 +1313,7082 @@
     1305
     1306  @Override
     1307  public boolean isSplittable() {
    -1308boolean result = isAvailable() 
     !hasReferences();
    -1309LOG.info("ASKED IF SPLITTABLE " + 
    result + " " + getRegionInfo().getShortNameToLog(),
    -1310  new Throwable("LOGGING: 
    REMOVE"));
    -1311// REMOVE BELOW
    -1312LOG.info("DEBUG LIST ALL FILES");
    -1313for (HStore store : 
    this.stores.values()) {
    -1314  LOG.info("store " + 
    store.getColumnFamilyName());
    -1315  for (HStoreFile sf : 
    store.getStorefiles()) {
    -1316
    LOG.info(sf.toStringDetailed());
    -1317  }
    -1318}
    -1319return result;
    -1320  }
    -1321
    -1322  @Override
    -1323  public boolean isMergeable() {
    -1324if (!isAvailable()) {
    -1325  LOG.debug("Region " + this
    -1326  + " is not mergeable because 
    it is closing or closed");
    -1327  return false;
    -1328}
    -1329if (hasReferences()) {
    -1330  LOG.debug("Region " + this
    -1331  + " is not mergeable because 
    it has references");
    -1332  return false;
    -1333}
    -1334
    -1335return true;
    +1308return isAvailable()  
    !hasReferences();
    +1309  }
    +1310
    +1311  @Override
    +1312  public boolean isMergeable() {
    +1313if (!isAvailable()) {
    +1314  LOG.debug("Region " + this
    +1315  + " is not mergeable because 
    it is closing or closed");
    +1316  return false;
    +1317}
    +1318if (hasReferences()) {
    +1319  LOG.debug("Region " + this
    +1320  + " is not mergeable because 
    it has references");
    +1321  return false;
    +1322}
    +1323
    +1324return true;
    +1325  }
    +1326
    +1327  public boolean areWritesEnabled() {
    +1328synchronized(this.writestate) {
    +1329  return 
    this.writestate.writesEnabled;
    +1330}
    +1331  }
    +1332
    +1333  @VisibleForTesting
    +1334  public MultiVersionConcurrencyControl 
    getMVCC() {
    +1335return mvcc;
     1336  }
     1337
    -1338  public boolean areWritesEnabled() {
    -1339synchronized(this.writestate) {
    -1340  return 
    this.writestate.writesEnabled;
    -1341}
    -1342  }
    -1343
    -1344  @VisibleForTesting
    -1345  public MultiVersionConcurrencyControl 
    getMVCC() {
    -1346return mvcc;
    -1347  }
    -1348
    -1349  @Override
    -1350  public long getMaxFlushedSeqId() {
    -1351return maxFlushedSeqId;
    +1338  @Override
    +1339  public long getMaxFlushedSeqId() {
    +1340return maxFlushedSeqId;
    +1341  }
    +1342
    +1343  /**
    +1344   * @return readpoint considering given 
    IsolationLevel. Pass {@code null} for default
    +1345   */
    +1346  public long 
    getReadPoint(IsolationLevel isolationLevel) {
    +1347if (isolationLevel != null 
     isolationLevel == IsolationLevel.READ_UNCOMMITTED) {
    +1348  // This scan can read even 
    uncommitted transactions
    +1349  return Long.MAX_VALUE;
    +1350}
    +1351return mvcc.getReadPoint();
     1352  }
     1353
    -1354  /**
    -1355   * @return readpoint considering given 
    IsolationLevel. Pass {@code null} for default
    -1356   */
    -1357  public long 
    getReadPoint(IsolationLevel isolationLevel) {
    -1358if (isolationLevel != null 
     isolationLevel == IsolationLevel.READ_UNCOMMITTED) {
    -1359  // This scan can read even 
    uncommitted transactions
    -1360  return Long.MAX_VALUE;
    -1361}
    -1362return mvcc.getReadPoint();
    -1363  }
    -1364
    -1365  public boolean 
    isLoadingCfsOnDemandDefault() {
    -1366return 
    this.isLoadingCfsOnDemandDefault;
    -1367  }
    -1368
    -1369  /**
    -1370   * Close down this HRegion.  Flush the 
    cache, shut down each HStore, don't
    -1371   * service any more calls.
    -1372   *
    -1373   * pThis method could take 
    some time to execute, so don't call it from a
    -1374   * time-sensitive thread.
    -1375   *
    -1376   * @return Vector of all the storage 
    files that the HRegion's component
    -1377   * HStores make use of.  It's a list 
    of all StoreFile objects. Returns empty
    -1378   * vector if already closed and null 
    if judged that it should not close.
    -1379   *
    -1380   * @throws IOException e
    -1381   * @throws DroppedSnapshotException 
    Thrown when replay of wal is required
    -1382   * because a Snapshot was not properly 
    persisted. The region is put in closing mode, and the
    -1383   * caller MUST abort after this.
    -1384   */
    -1385  public Mapbyte[], 
    ListHStoreFile close() throws IOException {
    -1386return close(false);
    -1387  }
    -1388
    -1389  private final Object closeLock = new 
    Object();
    -1390
    -1391  /** Conf key for the periodic flush 
    

      1   2   3   >