[27/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.

2019-01-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html 
b/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html
index 9913946..055a934 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html
@@ -236,18 +236,22 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
  
 
 
-org.apache.hadoop.hbase.tool
+org.apache.hadoop.hbase.thrift2.client
  
 
 
-org.apache.hadoop.hbase.util
+org.apache.hadoop.hbase.tool
  
 
 
-org.apache.hadoop.hbase.wal
+org.apache.hadoop.hbase.util
  
 
 
+org.apache.hadoop.hbase.wal
+ 
+
+
 org.apache.hadoop.hbase.zookeeper
  
 
@@ -1948,6 +1952,40 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
+
+
+
+Uses of Pair in org.apache.hadoop.hbase.thrift2.client
+
+Methods in org.apache.hadoop.hbase.thrift2.client
 that return Pair 
+
+Modifier and Type
+Method and Description
+
+
+
+PairInteger,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true";
 title="class or interface in java.lang">Integer>
+ThriftAdmin.getAlterStatus(byte[] tableName) 
+
+
+PairInteger,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true";
 title="class or interface in java.lang">Integer>
+ThriftAdmin.getAlterStatus(TableName tableName) 
+
+
+Pair
+ThriftConnection.DefaultThriftClientBuilder.getClient() 
+
+
+Pair
+ThriftConnection.HTTPThriftClientBuilder.getClient() 
+
+
+abstract Pair
+ThriftClientBuilder.getClient() 
+
+
+
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
index 25ec8fe..4e1d4c5 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
@@ -541,14 +541,14 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.util.PrettyPrinter.Unit
+org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
+org.apache.hadoop.hbase.util.ChecksumType
 org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE
+org.apache.hadoop.hbase.util.PrettyPrinter.Unit
+org.apache.hadoop.hbase.util.PoolMap.PoolType
+org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer
 (implements org.apache.hadoop.hbase.util.Bytes.Comparer)
 org.apache.hadoop.hbase.util.Order
 org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer
 (implements org.apache.hadoop.hbase.util.Bytes.Comparer)
-org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer
 (implements org.apache.hadoop.hbase.util.Bytes.Comparer)
-org.apache.hadoop.hbase.util.PoolMap.PoolType
-org.apache.hadoop.hbase.util.ChecksumType
-org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/util/package-use.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-use.html 
b/devapidocs/org/apache/hadoop/hbase/util/package-use.html
index 6336a97..7c797bf 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/package-use.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/package-use.html
@@ -311,14 +311,18 @@ service.
 
 
 
-org.apache.hadoop.hbase.tool
+org.apache.hadoop.hbase.thrift2.client
  
 
 
-org.apache.hadoop.hbase.tool.coprocessor
+org.apache.hadoop.hbase.tool
  
 
 
+org.apache.hadoop.hbase.tool.coprocessor
+ 
+
+
 org.apache.hadoop.hbase.types
 
 
@@ -326,19 +330,

[27/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.

2019-01-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
index ff0e4d5..c3c0cb1 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
@@ -761,58 +761,65 @@
 
 
 default void
+MasterObserver.postIsRpcThrottleEnabled(ObserverContext ctx,
+boolean rpcThrottleEnabled)
+Called after getting if is rpc throttle enabled.
+
+
+
+default void
 MasterObserver.postListDecommissionedRegionServers(ObserverContext ctx)
 Called after list decommissioned region servers.
 
 
-
+
 default void
 MasterObserver.postListNamespaceDescriptors(ObserverContext ctx,
 https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List descriptors)
 Called after a listNamespaceDescriptors request has been 
processed.
 
 
-
+
 default void
 MasterObserver.postListReplicationPeers(ObserverContext ctx,
 https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String regex)
 Called after list replication peers.
 
 
-
+
 default void
 MasterObserver.postListRSGroups(ObserverContext ctx)
 Called after listing region server group information.
 
 
-
+
 default void
 MasterObserver.postListSnapshot(ObserverContext ctx,
 SnapshotDescription snapshot)
 Called after listSnapshots request has been processed.
 
 
-
+
 default void
 MasterObserver.postLockHeartbeat(ObserverContext ctx)
 Called after heartbeat to a lock.
 
 
-
+
 default void
 RegionObserver.postMemStoreCompaction(ObserverContext c,
   Store store)
 Called after the in memory compaction is finished.
 
 
-
+
 default void
 MasterObserver.postMergeRegions(ObserverContext c,
 RegionInfo[] regionsToMerge)
 called after merge regions request.
 
 
-
+
 default void
 MasterObserver.postMergeRegionsCommitAction(ObserverContext ctx,
 RegionInfo[] regionsToMerge,
@@ -820,7 +827,7 @@
 This will be called after META step as part of regions 
merge transaction.
 
 
-
+
 default void
 MasterObserver.postModifyNamespace(ObserverContext ctx,
NamespaceDescriptor currentNsDescriptor)
@@ -829,7 +836,7 @@
 
 
 
-
+
 default void
 MasterObserver.postModifyNamespace(ObserverContext ctx,
NamespaceDescriptor oldNsDescriptor,
@@ -837,7 +844,7 @@
 Called after the modifyNamespace operation has been 
requested.
 
 
-
+
 default void
 MasterObserver.postModifyTable(ObserverContext ctx,
TableName tableName,
@@ -847,7 +854,7 @@
 
 
 
-
+
 default void
 MasterObserver.postModifyTable(ObserverContext ctx,
TableName tableName,
@@ -856,7 +863,7 @@
 Called after the modifyTable operation has been 
requested.
 
 
-
+
 default void
 MasterObserver.postMove(ObserverContext ctx,
 RegionInfo region,
@@ -865,7 +872,7 @@
 Called after the region move has been requested.
 
 
-
+
 default void
 MasterObserver.postMoveServers(ObserverContext ctx,
https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true";
 title="class or interface in java.util">Set
 servers, @@ -873,7 +880,7 @@ Called after servers are moved to target region server group - + default void MasterObserver.postMoveServersAndTables(ObserverContext ctx, https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true"; title="class or interface in java.util">Set
 servers, @@ -882,7 +889,7 @@ Called after servers are moved to target region server group - + default void MasterObserver.postMoveTables(ObserverContext ctx, https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true"; title="class or interface in java.util">Set tables, @@ -890,7 +897,7 @@ Called after servers are moved to target region server group - + default Cell RegionObserver.postMutationBeforeWAL

[27/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.

2019-01-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
index cd0ff28..4f9947f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
@@ -31,161 +31,161 @@
 023import static 
org.apache.hadoop.hbase.client.ConnectionUtils.resetController;
 024import static 
org.apache.hadoop.hbase.client.ConnectionUtils.translateException;
 025import static 
org.apache.hadoop.hbase.util.CollectionUtils.computeIfAbsent;
-026
-027import 
org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer;
-028
-029import java.io.IOException;
-030import java.util.ArrayList;
-031import java.util.Collections;
-032import java.util.HashMap;
-033import java.util.IdentityHashMap;
-034import java.util.List;
-035import java.util.Map;
-036import java.util.Optional;
-037import 
java.util.concurrent.CompletableFuture;
-038import 
java.util.concurrent.ConcurrentHashMap;
-039import 
java.util.concurrent.ConcurrentLinkedQueue;
-040import 
java.util.concurrent.ConcurrentMap;
-041import 
java.util.concurrent.ConcurrentSkipListMap;
-042import java.util.concurrent.TimeUnit;
-043import java.util.function.Supplier;
-044import java.util.stream.Collectors;
-045import java.util.stream.Stream;
-046
-047import 
org.apache.hadoop.hbase.CellScannable;
-048import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-049import 
org.apache.hadoop.hbase.HRegionLocation;
-050import 
org.apache.hadoop.hbase.ServerName;
-051import 
org.apache.hadoop.hbase.TableName;
-052import 
org.apache.yetus.audience.InterfaceAudience;
-053import org.slf4j.Logger;
-054import org.slf4j.LoggerFactory;
-055import 
org.apache.hadoop.hbase.client.MultiResponse.RegionResult;
-056import 
org.apache.hadoop.hbase.client.RetriesExhaustedException.ThrowableWithExtraContext;
-057import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-058import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-059import 
org.apache.hadoop.hbase.shaded.protobuf.ResponseConverter;
-060import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-061import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService;
-062import 
org.apache.hadoop.hbase.util.Bytes;
-063import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-064
-065/**
-066 * Retry caller for batch.
-067 * 

-068 * Notice that, the {@link #operationTimeoutNs} is the total time limit now which is the same with -069 * other single operations -070 *

-071 * And the {@link #maxAttempts} is a limit for each single operation in the batch logically. In the -072 * implementation, we will record a {@code tries} parameter for each operation group, and if it is -073 * split to several groups when retrying, the sub groups will inherit the {@code tries}. You can -074 * imagine that the whole retrying process is a tree, and the {@link #maxAttempts} is the limit of -075 * the depth of the tree. -076 */ -077@InterfaceAudience.Private -078class AsyncBatchRpcRetryingCaller { -079 -080 private static final Logger LOG = LoggerFactory.getLogger(AsyncBatchRpcRetryingCaller.class); -081 -082 private final HashedWheelTimer retryTimer; -083 -084 private final AsyncConnectionImpl conn; -085 -086 private final TableName tableName; -087 -088 private final List actions; -089 -090 private final List> futures; -091 -092 private final IdentityHashMap> action2Future; -093 -094 private final IdentityHashMap> action2Errors; -095 -096 private final long pauseNs; -097 -098 private final int maxAttempts; -099 -100 private final long operationTimeoutNs; -101 -102 private final long rpcTimeoutNs; -103 -104 private final int startLogErrorsCnt; -105 -106 private final long startNs; -107 -108 // we can not use HRegionLocation as the map key because the hashCode and equals method of -109 // HRegionLocation only consider serverName. -110 private static final class RegionRequest { -111 -112public final HRegionLocation loc; -113 -114public final ConcurrentLinkedQueue actions = new ConcurrentLinkedQueue<>(); -115 -116public RegionRequest(HRegionLocation loc) { -117 this.loc = loc; -118} -119 } -120 -121 private static final class ServerRequest { -122 -123public final ConcurrentMap actionsByRegion = -124new ConcurrentSkipListMap<>(Bytes.BYTES_COMPARATOR); -125 -126public void addAction(HRegionLocation loc, Action action) { -127 computeIfAbsent(actionsByRegion, loc


[27/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.

2019-01-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
index c4e8c8b..aa58108 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
@@ -82,602 +82,613 @@
 074  public static final String 
USER_COPROCESSORS_ENABLED_CONF_KEY =
 075"hbase.coprocessor.user.enabled";
 076  public static final boolean 
DEFAULT_USER_COPROCESSORS_ENABLED = true;
-077
-078  private static final Logger LOG = 
LoggerFactory.getLogger(CoprocessorHost.class);
-079  protected Abortable abortable;
-080  /** Ordered set of loaded coprocessors 
with lock */
-081  protected final SortedList 
coprocEnvironments =
-082  new SortedList<>(new 
EnvironmentPriorityComparator());
-083  protected Configuration conf;
-084  // unique file prefix to use for local 
copies of jars when classloading
-085  protected String pathPrefix;
-086  protected AtomicInteger loadSequence = 
new AtomicInteger();
-087
-088  public CoprocessorHost(Abortable 
abortable) {
-089this.abortable = abortable;
-090this.pathPrefix = 
UUID.randomUUID().toString();
-091  }
-092
-093  /**
-094   * Not to be confused with the 
per-object _coprocessors_ (above),
-095   * coprocessorNames is static and 
stores the set of all coprocessors ever
-096   * loaded by any thread in this JVM. It 
is strictly additive: coprocessors are
-097   * added to coprocessorNames, by 
checkAndLoadInstance() but are never removed, since
-098   * the intention is to preserve a 
history of all loaded coprocessors for
-099   * diagnosis in case of server crash 
(HBASE-4014).
-100   */
-101  private static Set 
coprocessorNames =
-102  Collections.synchronizedSet(new 
HashSet());
-103
-104  public static Set 
getLoadedCoprocessors() {
-105synchronized (coprocessorNames) {
-106  return new 
HashSet(coprocessorNames);
-107}
-108  }
-109
-110  /**
-111   * Used to create a parameter to the 
HServerLoad constructor so that
-112   * HServerLoad can provide information 
about the coprocessors loaded by this
-113   * regionserver.
-114   * (HBASE-4070: Improve region server 
metrics to report loaded coprocessors
-115   * to master).
-116   */
-117  public Set 
getCoprocessors() {
-118Set returnValue = new 
TreeSet<>();
-119for (E e: coprocEnvironments) {
-120  
returnValue.add(e.getInstance().getClass().getSimpleName());
-121}
-122return returnValue;
-123  }
-124
-125  /**
-126   * Load system coprocessors once only. 
Read the class names from configuration.
-127   * Called by constructor.
-128   */
-129  protected void 
loadSystemCoprocessors(Configuration conf, String confKey) {
-130boolean coprocessorsEnabled = 
conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
-131  DEFAULT_COPROCESSORS_ENABLED);
-132if (!coprocessorsEnabled) {
-133  return;
-134}
-135
-136Class implClass;
-137
-138// load default coprocessors from 
configure file
-139String[] defaultCPClasses = 
conf.getStrings(confKey);
-140if (defaultCPClasses == null || 
defaultCPClasses.length == 0)
-141  return;
-142
-143int priority = 
Coprocessor.PRIORITY_SYSTEM;
-144for (String className : 
defaultCPClasses) {
-145  className = className.trim();
-146  if (findCoprocessor(className) != 
null) {
-147// If already loaded will just 
continue
-148LOG.warn("Attempted duplicate 
loading of " + className + "; skipped");
-149continue;
-150  }
-151  ClassLoader cl = 
this.getClass().getClassLoader();
-152  
Thread.currentThread().setContextClassLoader(cl);
-153  try {
-154implClass = 
cl.loadClass(className);
-155// Add coprocessors as we go to 
guard against case where a coprocessor is specified twice
-156// in the configuration
-157E env = 
checkAndLoadInstance(implClass, priority, conf);
-158if (env != null) {
-159  
this.coprocEnvironments.add(env);
-160  LOG.info("System coprocessor {} 
loaded, priority={}.", className, priority);
-161  ++priority;
-162}
-163  } catch (Throwable t) {
-164// We always abort if system 
coprocessors cannot be loaded
-165abortServer(className, t);
-166  }
-167}
-168  }
-169
-170  /**
-171   * Load a coprocessor implementation 
into the host
-172   * @param path path to implementation 
jar
-173   * @param className the main class 
name
-174   * @param priority chaining priority
-175   * @param conf configuration 

[27/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.

2019-01-01 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html
index ee11a41..79ad648 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html
@@ -120,6 +120,6 @@
 
 
 
-Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
+Copyright © 2007–2019 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
index 85b5b1b..ae5fa8d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
@@ -191,6 +191,6 @@
 
 
 
-Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
+Copyright © 2007–2019 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
index 581a1bf..306c08d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
@@ -120,6 +120,6 @@
 
 
 
-Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
+Copyright © 2007–2019 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
index dd01782..cebfb9d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
@@ -162,6 +162,6 @@
 
 
 
-Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
+Copyright © 2007–2019 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
index c6dda63..f1d19b2 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
@@ -183,6 +183,6 @@
 
 
 
-Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
+Copyright © 2007–2019 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.html
 
b/devapidocs/org/apache/hadoop/hbase/

[27/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.

2018-12-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html
index 90e5a60..56edaab 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html
@@ -115,6 +115,13 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
  
 
 
+org.apache.hadoop.hbase.thrift2
+
+Provides an HBase http://thrift.apache.org/";>Thrift
+service.
+
+
+
 org.apache.hadoop.hbase.util
  
 
@@ -428,6 +435,37 @@ the order they are declared.
 
 
 
+
+
+
+Uses of BloomType in org.apache.hadoop.hbase.thrift2
+
+Methods in org.apache.hadoop.hbase.thrift2
 that return BloomType 
+
+Modifier and Type
+Method and Description
+
+
+
+static BloomType
+ThriftUtilities.bloomFilterFromThrift(org.apache.hadoop.hbase.thrift2.generated.TBloomFilterType in) 
+
+
+
+
+Methods in org.apache.hadoop.hbase.thrift2
 with parameters of type BloomType 
+
+Modifier and Type
+Method and Description
+
+
+
+static 
org.apache.hadoop.hbase.thrift2.generated.TBloomFilterType
+ThriftUtilities.bloomFilterFromHBase(BloomType in) 
+
+
+
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
index 3e892ef..9384f68 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
@@ -405,7 +405,7 @@
 HRegionServer.getRegionByEncodedName(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in 
java.lang">String encodedRegionName) 
 
 
-(package private) static HRegion
+static HRegion
 HRegion.newHRegion(org.apache.hadoop.fs.Path tableDir,
   WAL wal,
   org.apache.hadoop.fs.FileSystem fs,

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
index b06ddd9..bc92a6b 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
@@ -606,7 +606,7 @@
 
 
 
-(package private) static HRegion
+static HRegion
 HRegion.newHRegion(org.apache.hadoop.fs.Path tableDir,
   WAL wal,
   org.apache.hadoop.fs.FileSystem fs,

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index 8b58771..50c6f2d 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -716,20 +716,20 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
-org.apache.hadoop.hbase.regionserver.ScanType
-org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
-org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType
-org.apache.hadoop.hbase.regionserver.Region.Operation
-org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
 org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
 org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
-org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
+org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType
+org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type
 org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Re

[27/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.

2018-12-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
index b8d0e70..ac5997f 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HRegionServer.CompactionChecker
+private static class HRegionServer.CompactionChecker
 extends ScheduledChore
 
 
@@ -233,7 +233,7 @@ extends 
 
 instance
-private final HRegionServer instance
+private final HRegionServer instance
 
 
 
@@ -242,7 +242,7 @@ extends 
 
 majorCompactPriority
-private final int majorCompactPriority
+private final int majorCompactPriority
 
 
 
@@ -251,7 +251,7 @@ extends 
 
 DEFAULT_PRIORITY
-private static final int DEFAULT_PRIORITY
+private static final int DEFAULT_PRIORITY
 
 See Also:
 Constant
 Field Values
@@ -264,7 +264,7 @@ extends 
 
 iteration
-private long iteration
+private long iteration
 
 
 
@@ -281,7 +281,7 @@ extends 
 
 CompactionChecker
-CompactionChecker(HRegionServer h,
+CompactionChecker(HRegionServer h,
   int sleepTime,
   Stoppable stopper)
 
@@ -300,7 +300,7 @@ extends 
 
 chore
-protected void chore()
+protected void chore()
 Description copied from 
class: ScheduledChore
 The task to execute on each scheduled execution of the 
Chore
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
index 5d4def3..30395a4 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HRegionServer.MovedRegionInfo
+private static class HRegionServer.MovedRegionInfo
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 
 
@@ -218,7 +218,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 serverName
-private final ServerName serverName
+private final ServerName serverName
 
 
 
@@ -227,7 +227,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 seqNum
-private final long seqNum
+private final long seqNum
 
 
 
@@ -236,7 +236,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 ts
-private final long ts
+private final long ts
 
 
 
@@ -253,7 +253,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 MovedRegionInfo
-public MovedRegionInfo(ServerName serverName,
+public MovedRegionInfo(ServerName serverName,
long closeSeqNum)
 
 
@@ -271,7 +271,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getServerName
-public ServerName getServerName()
+public ServerName getServerName()
 
 
 
@@ -280,7 +280,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getSeqNum
-public long getSeqNum()
+public long getSeqNum()
 
 
 
@@ -289,7 +289,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getMoveTime
-public long getMoveTime()
+public long getMoveTime()
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
index 2a00361..fa924cd 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-protected static final class HRegionServer.MovedRegionsCleaner
+protected static final class HRegionServer.MovedRegionsCleaner
 extends ScheduledChore
 implements Stoppable
 Creates a Chore thread to clean the moved region 
cache.
@@ -242,7 +242,7 @@ implements 
 
 regionServer
-private HRegionServer regionServer
+private HRegionServer regionServer

[27/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.

2018-12-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
index ea05301..26a93dd 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
@@ -269,3590 +269,3574 @@
 261   */
 262  protected ClusterConnection 
clusterConnection;
 263
-264  /*
-265   * Long-living meta table locator, 
which is created when the server is started and stopped
-266   * when server shuts down. References 
to this locator shall be used to perform according
-267   * operations in EventHandlers. Primary 
reason for this decision is to make it mockable
-268   * for tests.
-269   */
-270  protected MetaTableLocator 
metaTableLocator;
-271
-272  /**
-273   * Go here to get table descriptors.
-274   */
-275  protected TableDescriptors 
tableDescriptors;
-276
-277  // Replication services. If no 
replication, this handler will be null.
-278  protected ReplicationSourceService 
replicationSourceHandler;
-279  protected ReplicationSinkService 
replicationSinkHandler;
-280
-281  // Compactions
-282  public CompactSplit 
compactSplitThread;
-283
-284  /**
-285   * Map of regions currently being 
served by this region server. Key is the
-286   * encoded region name.  All access 
should be synchronized.
-287   */
-288  protected final Map onlineRegions = new ConcurrentHashMap<>();
-289
-290  /**
-291   * Map of encoded region names to the 
DataNode locations they should be hosted on
-292   * We store the value as 
InetSocketAddress since this is used only in HDFS
-293   * API (create() that takes favored 
nodes as hints for placing file blocks).
-294   * We could have used ServerName here 
as the value class, but we'd need to
-295   * convert it to InetSocketAddress at 
some point before the HDFS API call, and
-296   * it seems a bit weird to store 
ServerName since ServerName refers to RegionServers
-297   * and here we really mean DataNode 
locations.
-298   */
-299  protected final Map regionFavoredNodesMap =
-300  new ConcurrentHashMap<>();
-301
-302  // Leases
-303  protected Leases leases;
+264  /**
+265   * Go here to get table descriptors.
+266   */
+267  protected TableDescriptors 
tableDescriptors;
+268
+269  // Replication services. If no 
replication, this handler will be null.
+270  protected ReplicationSourceService 
replicationSourceHandler;
+271  protected ReplicationSinkService 
replicationSinkHandler;
+272
+273  // Compactions
+274  public CompactSplit 
compactSplitThread;
+275
+276  /**
+277   * Map of regions currently being 
served by this region server. Key is the
+278   * encoded region name.  All access 
should be synchronized.
+279   */
+280  protected final Map onlineRegions = new ConcurrentHashMap<>();
+281
+282  /**
+283   * Map of encoded region names to the 
DataNode locations they should be hosted on
+284   * We store the value as 
InetSocketAddress since this is used only in HDFS
+285   * API (create() that takes favored 
nodes as hints for placing file blocks).
+286   * We could have used ServerName here 
as the value class, but we'd need to
+287   * convert it to InetSocketAddress at 
some point before the HDFS API call, and
+288   * it seems a bit weird to store 
ServerName since ServerName refers to RegionServers
+289   * and here we really mean DataNode 
locations.
+290   */
+291  protected final Map regionFavoredNodesMap =
+292  new ConcurrentHashMap<>();
+293
+294  // Leases
+295  protected Leases leases;
+296
+297  // Instance of the hbase executor 
executorService.
+298  protected ExecutorService 
executorService;
+299
+300  // If false, the file system has become 
unavailable
+301  protected volatile boolean fsOk;
+302  protected HFileSystem fs;
+303  protected HFileSystem walFs;
 304
-305  // Instance of the hbase executor 
executorService.
-306  protected ExecutorService 
executorService;
-307
-308  // If false, the file system has become 
unavailable
-309  protected volatile boolean fsOk;
-310  protected HFileSystem fs;
-311  protected HFileSystem walFs;
-312
-313  // Set when a report to the master 
comes back with a message asking us to
-314  // shutdown. Also set by call to stop 
when debugging or running unit tests
-315  // of HRegionServer in isolation.
-316  private volatile boolean stopped = 
false;
-317
-318  // Go down hard. Used if file system 
becomes unavailable and also in
-319  // debugging and unit tests.
-320  private volatile boolean 
abortRequested;
-321  public static f

[27/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.

2018-11-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html
deleted file mode 100644
index 908d0ce..000
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html
+++ /dev/null
@@ -1,165 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd";>
-
-
-
-
-
-Uses of Class 
org.apache.hadoop.hbase.master.assignment.ServerStateNode.ServerReportEvent 
(Apache HBase 3.0.0-SNAPSHOT API)
-
-
-
-
-
-
-
-JavaScript is disabled on your browser.
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-Prev
-Next
-
-
-Frames
-No Frames
-
-
-All Classes
-
-
-
-
-
-
-
-
-
-
-Uses of 
Classorg.apache.hadoop.hbase.master.assignment.ServerStateNode.ServerReportEvent
-
-
-
-
-
-Packages that use ServerStateNode.ServerReportEvent 
-
-Package
-Description
-
-
-
-org.apache.hadoop.hbase.master.assignment
- 
-
-
-
-
-
-
-
-
-
-Uses of ServerStateNode.ServerReportEvent
 in org.apache.hadoop.hbase.master.assignment
-
-Fields in org.apache.hadoop.hbase.master.assignment
 declared as ServerStateNode.ServerReportEvent 
-
-Modifier and Type
-Field and Description
-
-
-
-private ServerStateNode.ServerReportEvent
-ServerStateNode.reportEvent 
-
-
-
-
-
-
-
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-Prev
-Next
-
-
-Frames
-No Frames
-
-
-All Classes
-
-
-
-
-
-
-
-
-
-Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
-
-

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
index 44fe1fc..7e57352 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
@@ -159,10 +159,6 @@
 
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode state,
 long seqId) 
 
-
-protected void
-AssignmentManager.wakeServerReportEvent(ServerStateNode serverNode) 
-
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html
index 34c6cb8..03103e1 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html
@@ -37,7 +37,6 @@
 RegionStateStore
 RegionTransitionProcedure
 ServerStateNode
-ServerStateNode.ServerReportEvent
 SplitTableRegionProcedure
 TransitRegionStateProcedure
 UnassignProcedure

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
index 3235ced..577da39 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
@@ -210,22 +210,18 @@
 
 
 
-ServerStateNode.ServerReportEvent
- 
-
-
 SplitTableRegionProcedure
 
 The procedure to split a region i

[27/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.

2018-11-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html 
b/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html
index 44628e7..18eaf1f 100644
--- a/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html
+++ b/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html
@@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class RestoreSnapshotHelper
+public class RestoreSnapshotHelper
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 Helper to Restore/Clone a Snapshot
 
@@ -467,7 +467,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 LOG
-private static final org.slf4j.Logger LOG
+private static final org.slf4j.Logger LOG
 
 
 
@@ -476,7 +476,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 regionsMap
-private final https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
 title="class or interface in java.util">Map regionsMap
+private final https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
 title="class or interface in java.util">Map regionsMap
 
 
 
@@ -485,7 +485,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 parentsMap
-private final https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
 title="class or interface in java.util">MapString,PairString,https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String>> parentsMap
+private final https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
 title="class or interface in java.util">MapString,PairString,https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String>> parentsMap
 
 
 
@@ -494,7 +494,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 monitor
-private final ForeignExceptionDispatcher monitor
+private final ForeignExceptionDispatcher monitor
 
 
 
@@ -503,7 +503,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 status
-private final MonitoredTask status
+private final MonitoredTask status
 
 
 
@@ -512,7 +512,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 snapshotManifest
-private final SnapshotManifest snapshotManifest
+private final SnapshotManifest snapshotManifest
 
 
 
@@ -521,7 +521,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 snapshotDesc
-private 
final org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription
 snapshotDesc
+private 
final org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription
 snapshotDesc
 
 
 
@@ -530,7 +530,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 snapshotTable
-private final TableName snapshotTable
+private final TableName snapshotTable
 
 
 
@@ -539,7 +539,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 tableDesc
-private final TableDescriptor tableDesc
+private final TableDescriptor tableDesc
 
 
 
@@ -548,7 +548,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 rootDir
-private final org.apache.hadoop.fs.Path rootDir
+private final org.apache.hadoop.fs.Path rootDir
 
 
 
@@ -557,7 +557,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 tableDir
-private final org.apache.hadoop.fs.Path tableDir
+private final org.apache.hadoop.fs.Path tableDir
 
 
 
@@ -566,7 +566,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 conf
-private final org.apache.hadoop.conf.Configuration conf
+private final org.apache.hadoop.conf.Configuration conf
 
 
 
@@ -575,7 +575,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 fs
-private final org.apache.hadoop.fs.FileSystem fs
+private final org.apache.hadoop.fs.FileSystem fs
 
 
 
@@ -584,7 +584,7 @@ extends https://docs.oracle.com/javase

[27/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.

2018-10-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
index 0af8acd..c5f21ac 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
@@ -645,1615 +645,1597 @@
 637
proc.afterReplay(getEnvironment());
 638  }
 639});
-640
-641// 4. Push the procedures to the 
timeout executor
-642waitingTimeoutList.forEach(proc -> 
{
-643  
proc.afterReplay(getEnvironment());
-644  timeoutExecutor.add(proc);
-645});
-646// 5. restore locks
-647restoreLocks();
-648// 6. Push the procedure to the 
scheduler
-649
failedList.forEach(scheduler::addBack);
-650runnableList.forEach(p -> {
-651  p.afterReplay(getEnvironment());
-652  if (!p.hasParent()) {
-653
sendProcedureLoadedNotification(p.getProcId());
-654  }
-655  // If the procedure holds the lock, 
put the procedure in front
-656  // If its parent holds the lock, 
put the procedure in front
-657  // TODO. Is that possible that its 
ancestor holds the lock?
-658  // For now, the deepest procedure 
hierarchy is:
-659  // ModifyTableProcedure -> 
ReopenTableProcedure ->
-660  // MoveTableProcedure -> 
Unassign/AssignProcedure
-661  // But ModifyTableProcedure and 
ReopenTableProcedure won't hold the lock
-662  // So, check parent lock is 
enough(a tricky case is resovled by HBASE-21384).
-663  // If some one change or add new 
procedures making 'grandpa' procedure
-664  // holds the lock, but parent 
procedure don't hold the lock, there will
-665  // be a problem here. We have to 
check one procedure's ancestors.
-666  // And we need to change 
LockAndQueue.hasParentLock(Procedure proc) method
-667  // to check all ancestors too.
-668  if (p.isLockedWhenLoading() || 
(p.hasParent() && procedures
-669  
.get(p.getParentProcId()).isLockedWhenLoading())) {
-670scheduler.addFront(p, false);
-671  } else {
-672// if it was not, it can wait.
-673scheduler.addBack(p, false);
-674  }
-675});
-676// After all procedures put into the 
queue, signal the worker threads.
-677// Otherwise, there is a race 
condition. See HBASE-21364.
-678scheduler.signalAll();
-679  }
+640// 4. restore locks
+641restoreLocks();
+642
+643// 5. Push the procedures to the 
timeout executor
+644waitingTimeoutList.forEach(proc -> 
{
+645  
proc.afterReplay(getEnvironment());
+646  timeoutExecutor.add(proc);
+647});
+648
+649// 6. Push the procedure to the 
scheduler
+650
failedList.forEach(scheduler::addBack);
+651runnableList.forEach(p -> {
+652  p.afterReplay(getEnvironment());
+653  if (!p.hasParent()) {
+654
sendProcedureLoadedNotification(p.getProcId());
+655  }
+656  scheduler.addBack(p);
+657});
+658// After all procedures put into the 
queue, signal the worker threads.
+659// Otherwise, there is a race 
condition. See HBASE-21364.
+660scheduler.signalAll();
+661  }
+662
+663  /**
+664   * Initialize the procedure executor, 
but do not start workers. We will start them later.
+665   * 

+666 * It calls ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, and +667 * ensure a single executor, and start the procedure replay to resume and recover the previous +668 * pending and in-progress procedures. +669 * @param numThreads number of threads available for procedure execution. +670 * @param abortOnCorruption true if you want to abort your service in case a corrupted procedure +671 * is found on replay. otherwise false. +672 */ +673 public void init(int numThreads, boolean abortOnCorruption) throws IOException { +674// We have numThreads executor + one timer thread used for timing out +675// procedures and triggering periodic procedures. +676this.corePoolSize = numThreads; +677this.maxPoolSize = 10 * numThreads; +678LOG.info("Starting {} core workers (bigger of cpus/4 or 16) with max (burst) worker count={}", +679corePoolSize, maxPoolSize); 680 -681 /** -682 * Initialize the procedure executor, but do not start workers. We will start them later. -683 *

-684 * It calls ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, and -685 * ensure a single executor, and start the procedure replay to resume and recover the previous -686 * pending and in-progress procedures. -687 * @param numThreads number of threads available for p


[27/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.

2018-10-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
index c7d99b2..9d1542c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
@@ -382,1357 +382,1365 @@
 374for (int i = 0; i < 
this.curFunctionCosts.length; i++) {
 375  curFunctionCosts[i] = 
tempFunctionCosts[i];
 376}
-377LOG.info("start 
StochasticLoadBalancer.balancer, initCost=" + currentCost + ", functionCost="
-378+ functionCost());
+377double initCost = currentCost;
+378double newCost = currentCost;
 379
-380double initCost = currentCost;
-381double newCost = currentCost;
-382
-383long computedMaxSteps;
-384if (runMaxSteps) {
-385  computedMaxSteps = 
Math.max(this.maxSteps,
-386  ((long)cluster.numRegions * 
(long)this.stepsPerRegion * (long)cluster.numServers));
-387} else {
-388  computedMaxSteps = 
Math.min(this.maxSteps,
-389  ((long)cluster.numRegions * 
(long)this.stepsPerRegion * (long)cluster.numServers));
-390}
-391// Perform a stochastic walk to see 
if we can get a good fit.
-392long step;
-393
-394for (step = 0; step < 
computedMaxSteps; step++) {
-395  Cluster.Action action = 
nextAction(cluster);
-396
-397  if (action.type == Type.NULL) {
-398continue;
-399  }
-400
-401  cluster.doAction(action);
-402  updateCostsWithAction(cluster, 
action);
-403
-404  newCost = computeCost(cluster, 
currentCost);
-405
-406  // Should this be kept?
-407  if (newCost < currentCost) {
-408currentCost = newCost;
-409
-410// save for JMX
-411curOverallCost = currentCost;
-412for (int i = 0; i < 
this.curFunctionCosts.length; i++) {
-413  curFunctionCosts[i] = 
tempFunctionCosts[i];
-414}
-415  } else {
-416// Put things back the way they 
were before.
-417// TODO: undo by remembering old 
values
-418Action undoAction = 
action.undoAction();
-419cluster.doAction(undoAction);
-420updateCostsWithAction(cluster, 
undoAction);
-421  }
-422
-423  if 
(EnvironmentEdgeManager.currentTime() - startTime >
-424  maxRunningTime) {
-425break;
-426  }
-427}
-428long endTime = 
EnvironmentEdgeManager.currentTime();
-429
-430
metricsBalancer.balanceCluster(endTime - startTime);
-431
-432// update costs metrics
-433updateStochasticCosts(tableName, 
curOverallCost, curFunctionCosts);
-434if (initCost > currentCost) {
-435  plans = 
createRegionPlans(cluster);
-436  LOG.info("Finished computing new 
load balance plan. Computation took {}" +
-437" to try {} different iterations. 
 Found a solution that moves " +
-438"{} regions; Going from a 
computed cost of {}" +
-439" to a new cost of {}", 
java.time.Duration.ofMillis(endTime - startTime),
-440step, plans.size(), initCost, 
currentCost);
-441  return plans;
-442}
-443LOG.info("Could not find a better 
load balance plan.  Tried {} different configurations in " +
-444  "{}, and did not find anything with 
a computed cost less than {}", step,
-445  java.time.Duration.ofMillis(endTime 
- startTime), initCost);
-446return null;
-447  }
-448
-449  /**
-450   * update costs to JMX
-451   */
-452  private void 
updateStochasticCosts(TableName tableName, Double overall, Double[] subCosts) 
{
-453if (tableName == null) return;
-454
-455// check if the metricsBalancer is 
MetricsStochasticBalancer before casting
-456if (metricsBalancer instanceof 
MetricsStochasticBalancer) {
-457  MetricsStochasticBalancer balancer 
= (MetricsStochasticBalancer) metricsBalancer;
-458  // overall cost
-459  
balancer.updateStochasticCost(tableName.getNameAsString(),
-460"Overall", "Overall cost", 
overall);
-461
-462  // each cost function
-463  for (int i = 0; i < 
costFunctions.length; i++) {
-464CostFunction costFunction = 
costFunctions[i];
-465String costFunctionName = 
costFunction.getClass().getSimpleName();
-466Double costPercent = (overall == 
0) ? 0 : (subCosts[i] / overall);
-467// TODO: cost function may need a 
specific description
-468
balancer.updateStochasticCost(tableName.getNameAsString(), costFunctionName,
-469  "The percent of " + 
costFunctionName, costPercent);
-470  }
-471}
-472  }
-473
-474  private String funct

[27/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.

2018-10-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html
new file mode 100644
index 000..8ef0ff5
--- /dev/null
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html
@@ -0,0 +1,371 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+Source code
+
+
+
+
+001/**
+002 * Licensed to the Apache Software 
Foundation (ASF) under one
+003 * or more contributor license 
agreements.  See the NOTICE file
+004 * distributed with this work for 
additional information
+005 * regarding copyright ownership.  The 
ASF licenses this file
+006 * to you under the Apache License, 
Version 2.0 (the
+007 * "License"); you may not use this file 
except in compliance
+008 * with the License.  You may obtain a 
copy of the License at
+009 *
+010 * 
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or 
agreed to in writing, software
+013 * distributed under the License is 
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+015 * See the License for the specific 
language governing permissions and
+016 * limitations under the License.
+017 */
+018package 
org.apache.hadoop.hbase.procedure2.store.wal;
+019
+020import java.io.IOException;
+021import java.util.ArrayList;
+022import java.util.Collection;
+023import java.util.Collections;
+024import java.util.Comparator;
+025import java.util.HashMap;
+026import java.util.Iterator;
+027import java.util.List;
+028import java.util.Map;
+029import 
java.util.NoSuchElementException;
+030import 
org.apache.commons.lang3.mutable.MutableInt;
+031import 
org.apache.hadoop.hbase.procedure2.Procedure;
+032import 
org.apache.hadoop.hbase.procedure2.ProcedureUtil;
+033import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
+034import 
org.apache.yetus.audience.InterfaceAudience;
+035import org.slf4j.Logger;
+036import org.slf4j.LoggerFactory;
+037
+038import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos;
+039
+040/**
+041 * Used to build the tree for 
procedures.
+042 * 

+043 * We will group the procedures with the root procedure, and then validate each group. For each +044 * group of procedures(with the same root procedure), we will collect all the stack ids, if the max +045 * stack id is n, then all the stack ids should be from 0 to n, non-repetition and non-omission. If +046 * not, we will consider all the procedures in this group as corrupted. Please see the code in +047 * {@link #checkReady(Entry, Map)} method. +048 *

+049 * For the procedures not in any group, i.e, can not find the root procedure for these procedures, +050 * we will also consider them as corrupted. Please see the code in {@link #checkOrphan(Map)} method. +051 */ +052@InterfaceAudience.Private +053public final class WALProcedureTree { +054 +055 private static final Logger LOG = LoggerFactory.getLogger(WALProcedureTree.class); +056 +057 private static final class Entry { +058 +059private final ProcedureProtos.Procedure proc; +060 +061private final List subProcs = new ArrayList<>(); +062 +063public Entry(ProcedureProtos.Procedure proc) { +064 this.proc = proc; +065} +066 +067@Override +068public String toString() { +069 StringBuilder sb = new StringBuilder(); +070 sb.append("Procedure(pid="); +071 sb.append(proc.getProcId()); +072 sb.append(", ppid="); +073 sb.append(proc.hasParentId() ? proc.getParentId() : Procedure.NO_PROC_ID); +074 sb.append(", class="); +075 sb.append(proc.getClassName()); +076 sb.append(")"); +077 return sb.toString(); +078} +079 } +080 +081 // when loading we will iterator the procedures twice, so use this class to cache the deserialized +082 // result to prevent deserializing multiple times. +083 private static final class ProtoAndProc { +084private final ProcedureProtos.Procedure proto; +085 +086private Procedure proc; +087 +088public ProtoAndProc(ProcedureProtos.Procedure proto) { +089 this.proto = proto; +090} +091 +092public Procedure getProc() throws IOException { +093 if (proc == null) { +094proc = ProcedureUtil.convertToProcedure(proto); +095 } +096 return proc; +097} +098 } +099 +100 private final List validProcs = new ArrayList<>(); +101 +102 private final List corruptedProcs = new ArrayList<>(); +103 +104 private static boolean isFinished(ProcedureProtos.Procedure proc) { +105if (!proc.hasParentId()) { +106 switc


[27/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.

2018-10-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
 
b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
index 3d27554..fd63f54 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-static class StochasticLoadBalancer.RegionReplicaCandidateGenerator
+static class StochasticLoadBalancer.RegionReplicaCandidateGenerator
 extends StochasticLoadBalancer.CandidateGenerator
 Generates candidates which moves the replicas out of the 
region server for
  co-hosted region replicas
@@ -228,7 +228,7 @@ extends 
 
 randomGenerator
-StochasticLoadBalancer.RandomCandidateGenerator
 randomGenerator
+StochasticLoadBalancer.RandomCandidateGenerator
 randomGenerator
 
 
 
@@ -245,7 +245,7 @@ extends 
 
 RegionReplicaCandidateGenerator
-RegionReplicaCandidateGenerator()
+RegionReplicaCandidateGenerator()
 
 
 
@@ -262,7 +262,7 @@ extends 
 
 selectCoHostedRegionPerGroup
-int selectCoHostedRegionPerGroup(int[] primariesOfRegionsPerGroup,
+int selectCoHostedRegionPerGroup(int[] primariesOfRegionsPerGroup,
  int[] regionsPerGroup,
  int[] regionIndexToPrimaryIndex)
 Randomly select one regionIndex out of all region replicas 
co-hosted in the same group
@@ -284,7 +284,7 @@ extends 
 
 generate
-BaseLoadBalancer.Cluster.Action generate(BaseLoadBalancer.Cluster cluster)
+BaseLoadBalancer.Cluster.Action generate(BaseLoadBalancer.Cluster cluster)
 
 Specified by:
 generate in
 class StochasticLoadBalancer.CandidateGenerator

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
 
b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
index 1faf326..76e8473 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-static class StochasticLoadBalancer.RegionReplicaHostCostFunction
+static class StochasticLoadBalancer.RegionReplicaHostCostFunction
 extends StochasticLoadBalancer.CostFunction
 A cost function for region replicas. We give a very high 
cost to hosting
  replicas of the same region in the same host. We do not prevent the case
@@ -272,7 +272,7 @@ extends 
 
 REGION_REPLICA_HOST_COST_KEY
-private static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String REGION_REPLICA_HOST_COST_KEY
+private static final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String REGION_REPLICA_HOST_COST_KEY
 
 See Also:
 Constant
 Field Values
@@ -285,7 +285,7 @@ extends 
 
 DEFAULT_REGION_REPLICA_HOST_COST_KEY
-private static final float DEFAULT_REGION_REPLICA_HOST_COST_KEY
+private static final float DEFAULT_REGION_REPLICA_HOST_COST_KEY
 
 See Also:
 Constant
 Field Values
@@ -298,7 +298,7 @@ extends 
 
 maxCost
-long maxCost
+long maxCost
 
 
 
@@ -307,7 +307,7 @@ extends 
 
 costsPerGroup
-long[] costsPerGroup
+long[] costsPerGroup
 
 
 
@@ -316,7 +316,7 @@ extends 
 
 primariesOfRegionsPerGroup
-int[][] primariesOfRegionsPerGroup
+int[][] primariesOfRegionsPerGroup
 
 
 
@@ -333,7 +333,7 @@ extends 
 
 RegionReplicaHostCostFunction
-public RegionReplicaHostCostFunction(org.apache.hadoop.conf.Configuration conf)
+public RegionReplicaHostCostFunction(org.apache.hadoop.conf.Configuration conf)
 
 
 
@@ -350,7 +350,7 @@ extends 
 
 init
-void init(BaseLoadBalancer.Cluster cluster)
+void init(BaseLoadBalancer.Cluster cluster)
 Description copied from 
class: StochasticLoadBalancer.CostFunction
 Called once per LB invocation to give the cost function
  to initialize it's state, and perform any costly calculation.
@@ -366,7 +366,7 @@ extends 
 
 getMaxCost
-long getMaxCost(BaseLoadBalancer.Cluster cluster)
+long getMaxCost(BaseLoadBalancer.Cluster cluster)
 
 
 
@@ -375,7 +375,7 @@ extends 
 
 isNeeded

[27/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.

2018-10-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html
--
diff --git a/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html 
b/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html
index 2d9b60f..6af42fe 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html
@@ -146,8 +146,8 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.backup.TestBackupDeleteWithFailures.Failure
 org.apache.hadoop.hbase.backup.TestIncrementalBackupMergeWithFailures.FailurePhase
+org.apache.hadoop.hbase.backup.TestBackupDeleteWithFailures.Failure
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
index 06d362a..cb399cd 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
@@ -1167,8 +1167,16 @@
 
 
 static HBaseClassTestRule
+TestRestoreSnapshotFromClientGetCompactionState.CLASS_RULE 
+
+
+static HBaseClassTestRule
 TestMobCloneSnapshotFromClientNormal.CLASS_RULE 
 
+
+static HBaseClassTestRule
+TestMobRestoreSnapshotFromClientClone.CLASS_RULE 
+
 
 static HBaseClassTestRule
 TestCISleep.CLASS_RULE 
@@ -1223,43 +1231,43 @@
 
 
 static HBaseClassTestRule
-TestAsyncTableGetMultiThreaded.CLASS_RULE 
+TestMobRestoreSnapshotFromClientAfterSplittingRegions.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestSnapshotCloneIndependence.CLASS_RULE 
+TestAsyncTableGetMultiThreaded.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestCIDeleteRpcTimeout.CLASS_RULE 
+TestSnapshotCloneIndependence.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestZKAsyncRegistry.CLASS_RULE 
+TestCIDeleteRpcTimeout.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestAsyncProcedureAdminApi.CLASS_RULE 
+TestZKAsyncRegistry.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestAsyncClusterAdminApi2.CLASS_RULE 
+TestAsyncProcedureAdminApi.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestCIGetOperationTimeout.CLASS_RULE 
+TestAsyncClusterAdminApi2.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestAlwaysSetScannerId.CLASS_RULE 
+TestCIGetOperationTimeout.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestSnapshotFromClient.CLASS_RULE 
+TestAlwaysSetScannerId.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestRestoreSnapshotFromClient.CLASS_RULE 
+TestSnapshotFromClient.CLASS_RULE 
 
 
 static HBaseClassTestRule
@@ -1311,56 +1319,64 @@
 
 
 static HBaseClassTestRule
-TestBlockEvictionFromClient.CLASS_RULE 
+TestMobRestoreSnapshotFromClientGetCompactionState.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestCIBadHostname.CLASS_RULE 
+TestBlockEvictionFromClient.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE 
+TestCIBadHostname.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestAsyncTableAdminApi2.CLASS_RULE 
+TestCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestAsyncQuotaAdminApi.CLASS_RULE 
+TestAsyncTableAdminApi2.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestAsyncMetaRegionLocator.CLASS_RULE 
+TestAsyncQuotaAdminApi.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestAsyncTableNoncedRetry.CLASS_RULE 
+TestAsyncMetaRegionLocator.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestRegionLocationCaching.CLASS_RULE 
+TestAsyncTableNoncedRetry.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestConnectionUtils.CLASS_RULE 
+TestRegionLocationCaching.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestMobCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE 
+TestConnectionUtils.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestAsyncTableScanMetrics.CLASS_RULE 
+TestMobCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE 
 
 
 static HBaseClassTestRule
-TestFromClientSideScanExcpetionWithCoprocessor.CLASS_RULE 
+TestAsyncTableScanMetrics.CLASS_RULE 
 
 
 static HBaseClassTestRule
+TestFromClientSideScanExcpetionWithCoprocessor.CLASS_RULE 
+
+
+static HBaseClassTestRule
 TestMobCloneSnapshotFromClientError.CLASS_RULE 
 
+
+static HBaseClassTestRule
+TestRestoreSnapshotFromClientSchemaChange.CLASS_RULE 
+
 
 static HBaseC

[27/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.

2018-10-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
index 061ce80..bdfc3f8 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
@@ -39,2126 +39,2163 @@
 031import java.util.Set;
 032import 
java.util.concurrent.ConcurrentHashMap;
 033import 
java.util.concurrent.CopyOnWriteArrayList;
-034import java.util.concurrent.TimeUnit;
-035import 
java.util.concurrent.atomic.AtomicBoolean;
-036import 
java.util.concurrent.atomic.AtomicInteger;
-037import 
java.util.concurrent.atomic.AtomicLong;
-038import java.util.stream.Collectors;
-039import java.util.stream.Stream;
-040import 
org.apache.hadoop.conf.Configuration;
-041import 
org.apache.hadoop.hbase.HConstants;
-042import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
-043import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-044import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
-045import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
-046import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
-047import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
-048import 
org.apache.hadoop.hbase.security.User;
-049import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-050import 
org.apache.hadoop.hbase.util.IdLock;
-051import 
org.apache.hadoop.hbase.util.NonceKey;
-052import 
org.apache.hadoop.hbase.util.Threads;
-053import 
org.apache.yetus.audience.InterfaceAudience;
-054import org.slf4j.Logger;
-055import org.slf4j.LoggerFactory;
-056
-057import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-058import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+034import java.util.concurrent.Executor;
+035import java.util.concurrent.Executors;
+036import java.util.concurrent.TimeUnit;
+037import 
java.util.concurrent.atomic.AtomicBoolean;
+038import 
java.util.concurrent.atomic.AtomicInteger;
+039import 
java.util.concurrent.atomic.AtomicLong;
+040import java.util.stream.Collectors;
+041import java.util.stream.Stream;
+042import 
org.apache.hadoop.conf.Configuration;
+043import 
org.apache.hadoop.hbase.HConstants;
+044import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
+045import 
org.apache.hadoop.hbase.log.HBaseMarkers;
+046import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
+047import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
+048import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
+049import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureStoreListener;
+050import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
+051import 
org.apache.hadoop.hbase.security.User;
+052import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+053import 
org.apache.hadoop.hbase.util.IdLock;
+054import 
org.apache.hadoop.hbase.util.NonceKey;
+055import 
org.apache.hadoop.hbase.util.Threads;
+056import 
org.apache.yetus.audience.InterfaceAudience;
+057import org.slf4j.Logger;
+058import org.slf4j.LoggerFactory;
 059
-060import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
-061
-062/**
-063 * Thread Pool that executes the 
submitted procedures.
-064 * The executor has a ProcedureStore 
associated.
-065 * Each operation is logged and on 
restart the pending procedures are resumed.
-066 *
-067 * Unless the Procedure code throws an 
error (e.g. invalid user input)
-068 * the procedure will complete (at some 
point in time), On restart the pending
-069 * procedures are resumed and the once 
failed will be rolledback.
+060import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+061import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+062import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
+063
+064import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
+065
+066/**
+067 * Thread Pool that executes the 
submitted procedures.
+068 * The executor has a ProcedureStore 
associated.
+069 * Each operation is logged and on 
restart the pending procedures are resumed.
 070 *
-071 * The user can add procedures to the 
executor via submitProcedure(proc)
-072 * check for the finished state via 
isFinished(procId)
-073 * and get the result via 
getResult(procId)
-074 */
-075@InterfaceAudience.Private
-076public class 
ProcedureExecutor {
-0

[27/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.

2018-09-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html
index c7f1e58..c74635e 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html
@@ -164,897 +164,923 @@
 156   */
 157  private volatile boolean bypass = 
false;
 158
-159  public boolean isBypass() {
-160return bypass;
-161  }
-162
-163  /**
-164   * set the bypass to true
-165   * Only called in {@link 
ProcedureExecutor#bypassProcedure(long, long, boolean)} for now,
-166   * DO NOT use this method alone, since 
we can't just bypass
-167   * one single procedure. We need to 
bypass its ancestor too. So making it package private
-168   */
-169  void bypass() {
-170this.bypass = true;
-171  }
+159  /**
+160   * Indicate whether we need to persist 
the procedure to ProcedureStore after execution. Default to
+161   * true, and the implementation can all 
{@link #skipPersistence()} to let the framework skip the
+162   * persistence of the procedure.
+163   * 

+164 * This is useful when the procedure is in error and you want to retry later. The retry interval +165 * and the number of retries are usually not critical so skip the persistence can save some +166 * resources, and also speed up the restart processing. +167 *

+168 * Notice that this value will be reset to true every time before execution. And when rolling back +169 * we do not test this value. +170 */ +171 private boolean persist = true; 172 -173 /** -174 * The main code of the procedure. It must be idempotent since execute() -175 * may be called multiple times in case of machine failure in the middle -176 * of the execution. -177 * @param env the environment passed to the ProcedureExecutor -178 * @return a set of sub-procedures to run or ourselves if there is more work to do or null if the -179 * procedure is done. -180 * @throws ProcedureYieldException the procedure will be added back to the queue and retried later. -181 * @throws InterruptedException the procedure will be added back to the queue and retried later. -182 * @throws ProcedureSuspendedException Signal to the executor that Procedure has suspended itself and -183 * has set itself up waiting for an external event to wake it back up again. -184 */ -185 protected abstract Procedure[] execute(TEnvironment env) -186throws ProcedureYieldException, ProcedureSuspendedException, InterruptedException; -187 -188 /** -189 * The code to undo what was done by the execute() code. -190 * It is called when the procedure or one of the sub-procedures failed or an -191 * abort was requested. It should cleanup all the resources created by -192 * the execute() call. The implementation must be idempotent since rollback() -193 * may be called multiple time in case of machine failure in the middle -194 * of the execution. -195 * @param env the environment passed to the ProcedureExecutor -196 * @throws IOException temporary failure, the rollback will retry later -197 * @throws InterruptedException the procedure will be added back to the queue and retried later -198 */ -199 protected abstract void rollback(TEnvironment env) -200throws IOException, InterruptedException; -201 -202 /** -203 * The abort() call is asynchronous and each procedure must decide how to deal -204 * with it, if they want to be abortable. The simplest implementation -205 * is to have an AtomicBoolean set in the abort() method and then the execute() -206 * will check if the abort flag is set or not. -207 * abort() may be called multiple times from the client, so the implementation -208 * must be idempotent. -209 * -210 *

NOTE: abort() is not like Thread.interrupt(). It is just a notification -211 * that allows the procedure implementor abort. -212 */ -213 protected abstract boolean abort(TEnvironment env); -214 -215 /** -216 * The user-level code of the procedure may have some state to -217 * persist (e.g. input arguments or current position in the processing state) to -218 * be able to resume on failure. -219 * @param serializer stores the serializable state -220 */ -221 protected abstract void serializeStateData(ProcedureStateSerializer serializer) -222throws IOException; -223 -224 /** -225 * Called on store load to allow the user to decode the previously serialized -226 * state. -227 * @param serializer contains the serialized state -228 */ -229 protected abstract void deserializeStateData(ProcedureStateSerializer serializer) -230throws IOException; -231 -232 /** -233 * The {@link #doAcquireL


[27/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.

2018-09-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
index 55d677f..aae23e1 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
@@ -212,6 +212,10 @@
 HStore.getRegionFileSystem() 
 
 
+(package private) HRegionFileSystem
+HRegion.getRegionWALFileSystem() 
+
+
 static HRegionFileSystem
 HRegionFileSystem.openRegionFromFileSystem(org.apache.hadoop.conf.Configuration conf,
 org.apache.hadoop.fs.FileSystem fs,

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html
index 4c2b975..2abca97 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html
@@ -203,7 +203,7 @@
 static boolean
 WALSplitter.splitLogFile(org.apache.hadoop.fs.Path walDir,
 org.apache.hadoop.fs.FileStatus logfile,
-org.apache.hadoop.fs.FileSystem fs,
+org.apache.hadoop.fs.FileSystem walFS,
 org.apache.hadoop.conf.Configuration conf,
 CancelableProgressable reporter,
 LastSequenceId idChecker,
@@ -224,7 +224,7 @@
 WALSplitter(WALFactory factory,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.Path walDir,
-   org.apache.hadoop.fs.FileSystem fs,
+   org.apache.hadoop.fs.FileSystem walFS,
LastSequenceId idChecker,
SplitLogWorkerCoordination splitLogWorkerCoordination) 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
index 2ba4882..d9b1b13 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class AssignRegionHandler
+public class AssignRegionHandler
 extends EventHandler
 Handles opening of a region on a region server.
  
@@ -157,6 +157,10 @@ extends regionInfo 
 
 
+private RetryCounter
+retryCounter 
+
+
 private TableDescriptor
 tableDesc 
 
@@ -266,7 +270,7 @@ extends 
 
 LOG
-private static final org.slf4j.Logger LOG
+private static final org.slf4j.Logger LOG
 
 
 
@@ -275,7 +279,7 @@ extends 
 
 regionInfo
-private final RegionInfo regionInfo
+private final RegionInfo regionInfo
 
 
 
@@ -284,16 +288,25 @@ extends 
 
 tableDesc
-private final TableDescriptor tableDesc
+private final TableDescriptor tableDesc
 
 
 
 
 
-
+
 
 masterSystemTime
-private final long masterSystemTime
+private final long masterSystemTime
+
+
+
+
+
+
+
+retryCounter
+private final RetryCounter retryCounter
 
 
 
@@ -310,7 +323,7 @@ extends 
 
 AssignRegionHandler
-public AssignRegionHandler(RegionServerServices server,
+public AssignRegionHandler(RegionServerServices server,
RegionInfo regionInfo,
@Nullable
TableDescriptor tableDesc,
@@ -332,7 +345,7 @@ extends 
 
 getServer
-private RegionServerServices getServer()
+private RegionServerServices getServer()
 
 
 
@@ -341,7 +354,7 @@ extends 
 
 cleanUpAndReportFailure
-private void cleanUpAndReportFailure(https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException error)
+private void cleanUpAndReportFailure(https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException error)
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 
 Throws:
@@ -355,7 +368,7 @@ extends 
 
 process
-public void process()
+public void process()
  throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-exter

[27/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.

2018-09-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
index 49f081b..33c9cc0 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
@@ -35,309 +35,328 @@
 027import 
java.util.concurrent.BlockingQueue;
 028import 
java.util.concurrent.ConcurrentHashMap;
 029import 
java.util.concurrent.ConcurrentMap;
-030import 
java.util.concurrent.LinkedBlockingQueue;
-031import 
java.util.concurrent.ThreadPoolExecutor;
-032import java.util.concurrent.TimeUnit;
-033import 
java.util.concurrent.atomic.AtomicLong;
-034
-035import 
org.apache.yetus.audience.InterfaceAudience;
-036import org.slf4j.Logger;
-037import org.slf4j.LoggerFactory;
-038import 
org.apache.hadoop.hbase.monitoring.ThreadMonitoring;
+030import java.util.concurrent.Executors;
+031import 
java.util.concurrent.LinkedBlockingQueue;
+032import 
java.util.concurrent.ThreadPoolExecutor;
+033import java.util.concurrent.TimeUnit;
+034import 
java.util.concurrent.atomic.AtomicLong;
+035import 
org.apache.hadoop.hbase.monitoring.ThreadMonitoring;
+036import 
org.apache.yetus.audience.InterfaceAudience;
+037import org.slf4j.Logger;
+038import org.slf4j.LoggerFactory;
 039
 040import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
 041import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
 042import 
org.apache.hbase.thirdparty.com.google.common.collect.Maps;
-043import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-044
-045/**
-046 * This is a generic executor service. 
This component abstracts a
-047 * threadpool, a queue to which {@link 
EventType}s can be submitted,
-048 * and a 
Runnable that handles the object that is added to the 
queue.
-049 *
-050 * 

In order to create a new service, create an instance of this class and -051 * then do: instance.startExecutorService("myService");. When done -052 * call {@link #shutdown()}. -053 * -054 *

In order to use the service created above, call -055 * {@link #submit(EventHandler)}. -056 */ -057@InterfaceAudience.Private -058public class ExecutorService { -059 private static final Logger LOG = LoggerFactory.getLogger(ExecutorService.class); -060 -061 // hold the all the executors created in a map addressable by their names -062 private final ConcurrentHashMap executorMap = new ConcurrentHashMap<>(); +043import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ListenableFuture; +044import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ListeningScheduledExecutorService; +045import org.apache.hbase.thirdparty.com.google.common.util.concurrent.MoreExecutors; +046import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; +047 +048/** +049 * This is a generic executor service. This component abstracts a +050 * threadpool, a queue to which {@link EventType}s can be submitted, +051 * and a Runnable that handles the object that is added to the queue. +052 * +053 *

In order to create a new service, create an instance of this class and +054 * then do: instance.startExecutorService("myService");. When done +055 * call {@link #shutdown()}. +056 * +057 *

In order to use the service created above, call +058 * {@link #submit(EventHandler)}. +059 */ +060@InterfaceAudience.Private +061public class ExecutorService { +062 private static final Logger LOG = LoggerFactory.getLogger(ExecutorService.class); 063 -064 // Name of the server hosting this executor service. -065 private final String servername; +064 // hold the all the executors created in a map addressable by their names +065 private final ConcurrentHashMap executorMap = new ConcurrentHashMap<>(); 066 -067 /** -068 * Default constructor. -069 * @param servername Name of the hosting server. -070 */ -071 public ExecutorService(final String servername) { -072super(); -073this.servername = servername; -074 } -075 -076 /** -077 * Start an executor service with a given name. If there was a service already -078 * started with the same name, this throws a RuntimeException. -079 * @param name Name of the service to start. -080 */ -081 @VisibleForTesting -082 public void startExecutorService(String name, int maxThreads) { -083if (this.executorMap.get(name) != null) { -084 throw new RuntimeException("An executor service with the name " + name + -085


[27/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.

2018-09-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
index a5789e0..93a57cb 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
@@ -238,4120 +238,4119 @@
 230 * @see Admin
 231 */
 232@InterfaceAudience.Private
-233@InterfaceStability.Evolving
-234public class HBaseAdmin implements Admin 
{
-235  private static final Logger LOG = 
LoggerFactory.getLogger(HBaseAdmin.class);
-236
-237  private ClusterConnection connection;
-238
-239  private final Configuration conf;
-240  private final long pause;
-241  private final int numRetries;
-242  private final int syncWaitTimeout;
-243  private boolean aborted;
-244  private int operationTimeout;
-245  private int rpcTimeout;
-246
-247  private RpcRetryingCallerFactory 
rpcCallerFactory;
-248  private RpcControllerFactory 
rpcControllerFactory;
-249
-250  private NonceGenerator ng;
-251
-252  @Override
-253  public int getOperationTimeout() {
-254return operationTimeout;
-255  }
-256
-257  HBaseAdmin(ClusterConnection 
connection) throws IOException {
-258this.conf = 
connection.getConfiguration();
-259this.connection = connection;
-260
-261// TODO: receive 
ConnectionConfiguration here rather than re-parsing these configs every time.
-262this.pause = 
this.conf.getLong(HConstants.HBASE_CLIENT_PAUSE,
-263
HConstants.DEFAULT_HBASE_CLIENT_PAUSE);
-264this.numRetries = 
this.conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER,
-265
HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER);
-266this.operationTimeout = 
this.conf.getInt(HConstants.HBASE_CLIENT_OPERATION_TIMEOUT,
-267
HConstants.DEFAULT_HBASE_CLIENT_OPERATION_TIMEOUT);
-268this.rpcTimeout = 
this.conf.getInt(HConstants.HBASE_RPC_TIMEOUT_KEY,
-269
HConstants.DEFAULT_HBASE_RPC_TIMEOUT);
-270this.syncWaitTimeout = 
this.conf.getInt(
-271  
"hbase.client.sync.wait.timeout.msec", 10 * 6); // 10min
-272
-273this.rpcCallerFactory = 
connection.getRpcRetryingCallerFactory();
-274this.rpcControllerFactory = 
connection.getRpcControllerFactory();
-275
-276this.ng = 
this.connection.getNonceGenerator();
-277  }
-278
-279  @Override
-280  public void abort(String why, Throwable 
e) {
-281// Currently does nothing but throw 
the passed message and exception
-282this.aborted = true;
-283throw new RuntimeException(why, e);
-284  }
-285
-286  @Override
-287  public boolean isAborted() {
-288return this.aborted;
-289  }
-290
-291  @Override
-292  public boolean abortProcedure(final 
long procId, final boolean mayInterruptIfRunning)
-293  throws IOException {
-294return 
get(abortProcedureAsync(procId, mayInterruptIfRunning), this.syncWaitTimeout,
-295  TimeUnit.MILLISECONDS);
-296  }
-297
-298  @Override
-299  public Future 
abortProcedureAsync(final long procId, final boolean mayInterruptIfRunning)
-300  throws IOException {
-301Boolean abortProcResponse =
-302executeCallable(new 
MasterCallable(getConnection(),
-303getRpcControllerFactory()) 
{
-304  @Override
-305  protected AbortProcedureResponse 
rpcCall() throws Exception {
-306AbortProcedureRequest 
abortProcRequest =
-307
AbortProcedureRequest.newBuilder().setProcId(procId).build();
-308return 
master.abortProcedure(getRpcController(), abortProcRequest);
-309  }
-310}).getIsProcedureAborted();
-311return new AbortProcedureFuture(this, 
procId, abortProcResponse);
-312  }
-313
-314  @Override
-315  public List 
listTableDescriptors() throws IOException {
-316return 
listTableDescriptors((Pattern)null, false);
-317  }
-318
-319  @Override
-320  public List 
listTableDescriptors(Pattern pattern) throws IOException {
-321return listTableDescriptors(pattern, 
false);
-322  }
-323
-324  @Override
-325  public List 
listTableDescriptors(Pattern pattern, boolean includeSysTables)
-326  throws IOException {
-327return executeCallable(new 
MasterCallable>(getConnection(),
-328getRpcControllerFactory()) {
-329  @Override
-330  protected 
List rpcCall() throws Exception {
-331GetTableDescriptorsRequest req 
=
-332
RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables);
-333return 
ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(),
-334req));
-335  }
-336});
-337  }
-338
-339  

[27/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.

2018-09-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
index 2036bc9..bcfe1b9 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
@@ -28,113 +28,124 @@
 020
 021import java.io.IOException;
 022import java.util.ArrayList;
-023
-024import org.apache.hadoop.hbase.Cell;
-025import 
org.apache.yetus.audience.InterfaceAudience;
-026import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-027import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-028
-029import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
-030import 
org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException;
-031
-032/**
-033 * A filter that will only return the 
first KV from each row.
-034 * 

-035 * This filter can be used to more efficiently perform row count operations. -036 */ -037@InterfaceAudience.Public -038public class FirstKeyOnlyFilter extends FilterBase { -039 private boolean foundKV = false; -040 -041 public FirstKeyOnlyFilter() { -042 } -043 -044 @Override -045 public void reset() { -046foundKV = false; -047 } -048 -049 @Override -050 public boolean filterRowKey(Cell cell) throws IOException { -051// Impl in FilterBase might do unnecessary copy for Off heap backed Cells. -052return false; -053 } -054 -055 @Deprecated -056 @Override -057 public ReturnCode filterKeyValue(final Cell c) { -058return filterCell(c); -059 } -060 -061 @Override -062 public ReturnCode filterCell(final Cell c) { -063if(foundKV) return ReturnCode.NEXT_ROW; -064foundKV = true; -065return ReturnCode.INCLUDE; -066 } -067 -068 public static Filter createFilterFromArguments(ArrayList filterArguments) { -069 Preconditions.checkArgument(filterArguments.isEmpty(), -070"Expected 0 but got: %s", filterArguments.size()); -071return new FirstKeyOnlyFilter(); -072 } -073 -074 /** -075 * @return true if first KV has been found. -076 */ -077 protected boolean hasFoundKV() { -078return this.foundKV; -079 } -080 -081 /** -082 * -083 * @param value update {@link #foundKV} flag with value. -084 */ -085 protected void setFoundKV(boolean value) { -086this.foundKV = value; -087 } -088 -089 /** -090 * @return The filter serialized using pb -091 */ -092 @Override -093 public byte [] toByteArray() { -094 FilterProtos.FirstKeyOnlyFilter.Builder builder = -095 FilterProtos.FirstKeyOnlyFilter.newBuilder(); -096return builder.build().toByteArray(); -097 } -098 -099 /** -100 * @param pbBytes A pb serialized {@link FirstKeyOnlyFilter} instance -101 * @return An instance of {@link FirstKeyOnlyFilter} made from bytes -102 * @throws org.apache.hadoop.hbase.exceptions.DeserializationException -103 * @see #toByteArray -104 */ -105 public static FirstKeyOnlyFilter parseFrom(final byte [] pbBytes) -106 throws DeserializationException { -107// There is nothing to deserialize. Why do this at all? -108try { -109 FilterProtos.FirstKeyOnlyFilter.parseFrom(pbBytes); -110} catch (InvalidProtocolBufferException e) { -111 throw new DeserializationException(e); -112} -113// Just return a new instance. -114return new FirstKeyOnlyFilter(); -115 } -116 -117 /** -118 * @param o the other filter to compare with -119 * @return true if and only if the fields of the filter that are serialized -120 * are equal to the corresponding fields in other. Used for testing. -121 */ -122 @Override -123 boolean areSerializedFieldsEqual(Filter o) { -124if (o == this) return true; -125if (!(o instanceof FirstKeyOnlyFilter)) return false; -126 -127return true; -128 } -129} +023import java.util.Objects; +024 +025import org.apache.hadoop.hbase.Cell; +026import org.apache.yetus.audience.InterfaceAudience; +027import org.apache.hadoop.hbase.exceptions.DeserializationException; +028import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos; +029 +030import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +031import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException; +032 +033/** +034 * A filter that will only return the first KV from each row. +035 *

+036 * This filter can be used to more efficiently perform row count operations. +037 */ +038@InterfaceAudience.Public +039public class FirstKeyOnlyFilter extends FilterBase { +040 private boolean foundKV = false; +041 +042 public FirstKeyOnlyFilter() { +043 } +04


[27/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.

2018-08-31 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
index 71b9375..8f7d034 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private final class HStore.StoreFlusherImpl
+private final class HStore.StoreFlusherImpl
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 implements StoreFlushContext
 
@@ -279,7 +279,7 @@ implements 
 
 tracker
-private final FlushLifeCycleTracker tracker
+private final FlushLifeCycleTracker tracker
 
 
 
@@ -288,7 +288,7 @@ implements 
 
 cacheFlushSeqNum
-private final long cacheFlushSeqNum
+private final long cacheFlushSeqNum
 
 
 
@@ -297,7 +297,7 @@ implements 
 
 snapshot
-private MemStoreSnapshot snapshot
+private MemStoreSnapshot snapshot
 
 
 
@@ -306,7 +306,7 @@ implements 
 
 tempFiles
-private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List tempFiles
+private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List tempFiles
 
 
 
@@ -315,7 +315,7 @@ implements 
 
 committedFiles
-private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List committedFiles
+private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List committedFiles
 
 
 
@@ -324,7 +324,7 @@ implements 
 
 cacheFlushCount
-private long cacheFlushCount
+private long cacheFlushCount
 
 
 
@@ -333,7 +333,7 @@ implements 
 
 cacheFlushSize
-private long cacheFlushSize
+private long cacheFlushSize
 
 
 
@@ -342,7 +342,7 @@ implements 
 
 outputFileSize
-private long outputFileSize
+private long outputFileSize
 
 
 
@@ -359,7 +359,7 @@ implements 
 
 StoreFlusherImpl
-private StoreFlusherImpl(long cacheFlushSeqNum,
+private StoreFlusherImpl(long cacheFlushSeqNum,
  FlushLifeCycleTracker tracker)
 
 
@@ -377,7 +377,7 @@ implements 
 
 prepare
-public MemStoreSize prepare()
+public MemStoreSize prepare()
 This is not thread safe. The caller should have a lock on 
the region or the store.
  If necessary, the lock can be added with the patch provided in 
HBASE-10087
 
@@ -394,7 +394,7 @@ implements 
 
 flushCache
-public void flushCache(MonitoredTask status)
+public void flushCache(MonitoredTask status)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
interface: StoreFlushContext
 Flush the cache (create the new store file)
@@ -415,7 +415,7 @@ implements 
 
 commit
-public boolean commit(MonitoredTask status)
+public boolean commit(MonitoredTask status)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
interface: StoreFlushContext
 Commit the flush - add the store file to the store and 
clear the
@@ -440,7 +440,7 @@ implements 
 
 getOutputFileSize
-public long getOutputFileSize()
+public long getOutputFileSize()
 
 Specified by:
 getOutputFileSize in
 interface StoreFlushContext
@@ -455,7 +455,7 @@ implements 
 
 getCommittedFiles
-public https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List getCommittedFiles()
+public https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in 
java.util">List getCommittedFiles()
 Description copied from 
interface: StoreFlushContext
 Returns the newly committed files from the flush. Called 
only if commit returns true
 
@@ -472,7 +472,7 @@ implements 
 
 replayFlush
-public void replayFlush(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">ListString> fileNames,
+public void replayFlush(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List

[27/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.

2018-08-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
index e42e654..fe4e081 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
@@ -560,7 +560,7 @@
 552TransitRegionStateProcedure proc;
 553regionNode.lock();
 554try {
-555  preTransitCheck(regionNode, 
RegionStates.STATES_EXPECTED_ON_OPEN);
+555  preTransitCheck(regionNode, 
STATES_EXPECTED_ON_ASSIGN);
 556  proc = 
TransitRegionStateProcedure.assign(getProcedureEnvironment(), regionInfo, 
sn);
 557  regionNode.setProcedure(proc);
 558} finally {
@@ -581,7 +581,7 @@
 573TransitRegionStateProcedure proc;
 574regionNode.lock();
 575try {
-576  preTransitCheck(regionNode, 
RegionStates.STATES_EXPECTED_ON_CLOSE);
+576  preTransitCheck(regionNode, 
STATES_EXPECTED_ON_UNASSIGN_OR_MOVE);
 577  proc = 
TransitRegionStateProcedure.unassign(getProcedureEnvironment(), regionInfo);
 578  regionNode.setProcedure(proc);
 579} finally {
@@ -599,7 +599,7 @@
 591TransitRegionStateProcedure proc;
 592regionNode.lock();
 593try {
-594  preTransitCheck(regionNode, 
RegionStates.STATES_EXPECTED_ON_CLOSE);
+594  preTransitCheck(regionNode, 
STATES_EXPECTED_ON_UNASSIGN_OR_MOVE);
 595  regionNode.checkOnline();
 596  proc = 
TransitRegionStateProcedure.move(getProcedureEnvironment(), regionInfo, 
targetServer);
 597  regionNode.setProcedure(proc);
@@ -1419,448 +1419,483 @@
 1411  }
 1412
 1413  // 

-1414  //  Region Status update
-1415  //  Should only be called in 
TransitRegionStateProcedure
-1416  // 

-1417  private void 
transitStateAndUpdate(RegionStateNode regionNode, RegionState.State newState,
-1418  RegionState.State... 
expectedStates) throws IOException {
-1419RegionState.State state = 
regionNode.getState();
-1420regionNode.transitionState(newState, 
expectedStates);
-1421boolean succ = false;
-1422try {
-1423  
regionStateStore.updateRegionLocation(regionNode);
-1424  succ = true;
-1425} finally {
-1426  if (!succ) {
-1427// revert
-1428regionNode.setState(state);
-1429  }
-1430}
-1431  }
-1432
-1433  // should be called within the 
synchronized block of RegionStateNode
-1434  void regionOpening(RegionStateNode 
regionNode) throws IOException {
-1435transitStateAndUpdate(regionNode, 
State.OPENING, RegionStates.STATES_EXPECTED_ON_OPEN);
-1436
regionStates.addRegionToServer(regionNode);
-1437// update the operation count 
metrics
-1438
metrics.incrementOperationCounter();
-1439  }
-1440
-1441  // should be called within the 
synchronized block of RegionStateNode.
-1442  // The parameter 'giveUp' means 
whether we will try to open the region again, if it is true, then
-1443  // we will persist the FAILED_OPEN 
state into hbase:meta.
-1444  void regionFailedOpen(RegionStateNode 
regionNode, boolean giveUp) throws IOException {
-1445RegionState.State state = 
regionNode.getState();
-1446ServerName regionLocation = 
regionNode.getRegionLocation();
-1447if (giveUp) {
-1448  
regionNode.setState(State.FAILED_OPEN);
-1449  
regionNode.setRegionLocation(null);
-1450  boolean succ = false;
-1451  try {
-1452
regionStateStore.updateRegionLocation(regionNode);
-1453succ = true;
-1454  } finally {
-1455if (!succ) {
-1456  // revert
-1457  regionNode.setState(state);
-1458  
regionNode.setRegionLocation(regionLocation);
-1459}
-1460  }
-1461}
-1462if (regionLocation != null) {
-1463  
regionStates.removeRegionFromServer(regionLocation, regionNode);
-1464}
-1465  }
-1466
-1467  // should be called within the 
synchronized block of RegionStateNode
-1468  void regionOpened(RegionStateNode 
regionNode) throws IOException {
-1469// TODO: OPENING Updates hbase:meta 
too... we need to do both here and there?
-1470// That is a lot of hbase:meta 
writing.
-1471transitStateAndUpdate(regionNode, 
State.OPEN, RegionStates.STATES_EXPECTED_ON_OPEN);
-1472RegionInfo hri = 
regionNode.getRegionInfo();
-1473if (isMetaRegion(hri)) {
-1474  // Usually we'd set a table 
ENABLED at this stage but hbase:meta is ALWAYs enabled, it
-1475  // can't be disabled -- so skip 
the RPC (besides... enabled is managed by TableStateManager
-147

[27/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.

2018-08-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
index 2db4997..14fcebb 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
@@ -29,146 +29,161 @@
 021
 022import java.io.IOException;
 023import java.util.ArrayList;
-024
-025import org.apache.hadoop.hbase.Cell;
-026import 
org.apache.hadoop.hbase.CompareOperator;
-027import 
org.apache.yetus.audience.InterfaceAudience;
-028import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-029import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-030import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-031import 
org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException;
-032
-033/**
-034 * This filter is used to filter based on 
the key. It takes an operator
-035 * (equal, greater, not equal, etc) and a 
byte [] comparator for the row,
-036 * and column qualifier portions of a 
key.
-037 * 

-038 * This filter can be wrapped with {@link WhileMatchFilter} to add more control. -039 *

-040 * Multiple filters can be combined using {@link FilterList}. -041 *

-042 * If an already known row range needs to be scanned, -043 * use {@link org.apache.hadoop.hbase.CellScanner} start -044 * and stop rows directly rather than a filter. -045 */ -046@InterfaceAudience.Public -047public class RowFilter extends CompareFilter { -048 -049 private boolean filterOutRow = false; -050 -051 /** -052 * Constructor. -053 * @param rowCompareOp the compare op for row matching -054 * @param rowComparator the comparator for row matching -055 * @deprecated Since 2.0.0. Will remove in 3.0.0. Use -056 * {@link #RowFilter(CompareOperator, ByteArrayComparable)}} instead. -057 */ -058 @Deprecated -059 public RowFilter(final CompareOp rowCompareOp, -060 final ByteArrayComparable rowComparator) { -061super(rowCompareOp, rowComparator); -062 } -063 -064 /** -065 * Constructor. -066 * @param op the compare op for row matching -067 * @param rowComparator the comparator for row matching -068 */ -069 public RowFilter(final CompareOperator op, -070 final ByteArrayComparable rowComparator) { -071super(op, rowComparator); -072 } -073 -074 @Override -075 public void reset() { -076this.filterOutRow = false; -077 } -078 -079 @Deprecated -080 @Override -081 public ReturnCode filterKeyValue(final Cell c) { -082return filterCell(c); -083 } -084 -085 @Override -086 public ReturnCode filterCell(final Cell v) { -087if(this.filterOutRow) { -088 return ReturnCode.NEXT_ROW; -089} -090return ReturnCode.INCLUDE; -091 } -092 -093 @Override -094 public boolean filterRowKey(Cell firstRowCell) { -095if (compareRow(getCompareOperator(), this.comparator, firstRowCell)) { -096 this.filterOutRow = true; -097} -098return this.filterOutRow; -099 } -100 -101 @Override -102 public boolean filterRow() { -103return this.filterOutRow; -104 } -105 -106 public static Filter createFilterFromArguments(ArrayList filterArguments) { -107@SuppressWarnings("rawtypes") // for arguments -108ArrayList arguments = CompareFilter.extractArguments(filterArguments); -109CompareOperator compareOp = (CompareOperator)arguments.get(0); -110ByteArrayComparable comparator = (ByteArrayComparable)arguments.get(1); -111return new RowFilter(compareOp, comparator); -112 } -113 -114 /** -115 * @return The filter serialized using pb -116 */ -117 @Override -118 public byte [] toByteArray() { -119FilterProtos.RowFilter.Builder builder = -120 FilterProtos.RowFilter.newBuilder(); -121 builder.setCompareFilter(super.convert()); -122return builder.build().toByteArray(); -123 } -124 -125 /** -126 * @param pbBytes A pb serialized {@link RowFilter} instance -127 * @return An instance of {@link RowFilter} made from bytes -128 * @throws DeserializationException -129 * @see #toByteArray -130 */ -131 public static RowFilter parseFrom(final byte [] pbBytes) -132 throws DeserializationException { -133FilterProtos.RowFilter proto; -134try { -135 proto = FilterProtos.RowFilter.parseFrom(pbBytes); -136} catch (InvalidProtocolBufferException e) { -137 throw new DeserializationException(e); -138} -139final CompareOperator valueCompareOp = -140 CompareOperator.valueOf(proto.getCompareFilter().getCompareOp().name()); -141ByteArrayComparable valueComparator = null; -142try { -143 if (proto.getCompareFilter().hasComparator()) { -144valueComparator = Protobu


[27/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  Collection stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}

[27/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.

2018-08-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
index 8ffeb95..b1fbd27 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":6,"i5":10,"i6":10,"i7":10,"i8":10,"i9":6,"i10":10,"i11":10,"i12":10,"i13":10,"i14":6,"i15":10,"i16":10,"i17":10,"i18":6,"i19":6,"i20":10,"i21":10,"i22":10,"i23":6,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":6,"i32":10,"i33":10,"i34":6,"i35":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":42,"i1":42,"i2":38,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":38,"i10":38,"i11":38,"i12":42,"i13":42,"i14":42,"i15":38,"i16":42,"i17":42,"i18":42,"i19":42,"i20":38,"i21":42,"i22":42,"i23":38};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
 
 
 Prev Class
-Next Class
+Next Class
 
 
 Frames
@@ -121,58 +121,20 @@ var activeTableTab = "activeTableTab";
 AssignProcedure, UnassignProcedure
 
 
+Deprecated. 
+Do not use any 
more.
+
 
-@InterfaceAudience.Private
-public abstract class RegionTransitionProcedure
+https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
 title="class or interface in java.lang">@Deprecated
+ @InterfaceAudience.Private
+public abstract class RegionTransitionProcedure
 extends Procedure
 implements TableProcedureInterface, RemoteProcedureDispatcher.RemoteProcedure
-Base class for the Assign and Unassign Procedure.
-
- Locking:
- Takes exclusive lock on the region being assigned/unassigned. Thus, there can 
only be one
- RegionTransitionProcedure per region running at a time (see 
MasterProcedureScheduler).
-
- This procedure is asynchronous and responds to external events.
- The AssignmentManager will notify this procedure when the RS completes
- the operation and reports the transitioned state
- (see the Assign and Unassign class for more detail).
-
- Procedures move from the REGION_TRANSITION_QUEUE state when they are
- first submitted, to the REGION_TRANSITION_DISPATCH state when the request
- to remote server is sent and the Procedure is suspended waiting on external
- event to be woken again. Once the external event is triggered, Procedure
- moves to the REGION_TRANSITION_FINISH state.
-
- NOTE: AssignProcedure and 
UnassignProcedure 
should not be thought of
- as being asymmetric, at least currently.
- 
- AssignProcedure 
moves through all the above described states and implements methods
- associated with each while UnassignProcedure 
starts at state
- REGION_TRANSITION_DISPATCH and state REGION_TRANSITION_QUEUE is not 
supported.
-
- When any step in AssignProcedure 
fails, failure handler
- AssignProcedure#handleFailure(MasterProcedureEnv, RegionStateNode) 
re-attempts the
- assignment by setting the procedure state to REGION_TRANSITION_QUEUE and 
forces
- assignment to a different target server by setting AssignProcedure.forceNewPlan.
 When
- the number of attempts reaches threshold configuration 
'hbase.assignment.maximum.attempts',
- the procedure is aborted. For UnassignProcedure, 
similar re-attempts are
- intentionally not implemented. It is a 'one shot' procedure. See its class 
doc for how it
- handles failure.
- 
- If we find a region in an 'unexpected' state, we'll complain and retry 
with backoff forever.
- The 'unexpected' state needs to be fixed either by another running Procedure 
or by operator
- intervention (Regions in 'unexpected' state indicates bug or unexpected 
transition type).
- For this to work, subclasses need to persist the 'attempt' counter kept in 
this class when
- they do serializeStateData and restore it inside their deserializeStateData, 
just as they do
- for regionInfo.
- 
- 
- 
-
- TODO: Considering it is a priority doing all we can to get make a region 
available as soon as
- possible, re-attempting with any target makes sense if specified target fails 
in case of
- AssignProcedure. 
For UnassignProcedure, 
our concern is preventing data loss
- on failed unassign. See class doc for explanation.
+Leave here only for checking if we can successfully start 
the master.
+
+See Also:
+Tra

[27/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.

2018-08-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
index 39170f0..7859ebc 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
@@ -230,564 +230,567 @@
 222  }
 223}
 224  } catch (InterruptedException e) 
{
-225e.printStackTrace();
-226  }
-227}
-228
-229@Override
-230public void setup(Context context) 
throws IOException {
-231  cfRenameMap = 
createCfRenameMap(context.getConfiguration());
-232  filter = 
instantiateFilter(context.getConfiguration());
-233  int reduceNum = 
context.getNumReduceTasks();
-234  Configuration conf = 
context.getConfiguration();
-235  TableName tableName = 
TableName.valueOf(context.getConfiguration().get(TABLE_NAME));
-236  try (Connection conn = 
ConnectionFactory.createConnection(conf);
-237  RegionLocator regionLocator = 
conn.getRegionLocator(tableName)) {
-238byte[][] startKeys = 
regionLocator.getStartKeys();
-239if (startKeys.length != 
reduceNum) {
-240  throw new IOException("Region 
split after job initialization");
-241}
-242CellWritableComparable[] 
startKeyWraps =
-243new 
CellWritableComparable[startKeys.length - 1];
-244for (int i = 1; i < 
startKeys.length; ++i) {
-245  startKeyWraps[i - 1] =
-246  new 
CellWritableComparable(KeyValueUtil.createFirstOnRow(startKeys[i]));
-247}
-248
CellWritableComparablePartitioner.START_KEYS = startKeyWraps;
-249  }
-250}
-251  }
-252
-253  /**
-254   * A mapper that just writes out 
KeyValues.
-255   */
-256  
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="EQ_COMPARETO_USE_OBJECT_EQUALS",
-257  justification="Writables are going 
away and this has been this way forever")
-258  public static class CellImporter 
extends TableMapper {
-259private Map 
cfRenameMap;
-260private Filter filter;
-261private static final Logger LOG = 
LoggerFactory.getLogger(CellImporter.class);
-262
-263/**
-264 * @param row  The current table row 
key.
-265 * @param value  The columns.
-266 * @param context  The current 
context.
-267 * @throws IOException When something 
is broken with the data.
-268 */
-269@Override
-270public void 
map(ImmutableBytesWritable row, Result value,
-271  Context context)
-272throws IOException {
-273  try {
-274if (LOG.isTraceEnabled()) {
-275  LOG.trace("Considering the 
row."
-276  + Bytes.toString(row.get(), 
row.getOffset(), row.getLength()));
-277}
-278if (filter == null
-279|| 
!filter.filterRowKey(PrivateCellUtil.createFirstOnRow(row.get(), 
row.getOffset(),
-280(short) 
row.getLength( {
-281  for (Cell kv : 
value.rawCells()) {
-282kv = filterKv(filter, kv);
-283// skip if we filtered it 
out
-284if (kv == null) continue;
-285context.write(row, new 
MapReduceExtendedCell(convertKv(kv, cfRenameMap)));
-286  }
-287}
-288  } catch (InterruptedException e) 
{
-289e.printStackTrace();
-290  }
-291}
-292
-293@Override
-294public void setup(Context context) 
{
-295  cfRenameMap = 
createCfRenameMap(context.getConfiguration());
-296  filter = 
instantiateFilter(context.getConfiguration());
-297}
-298  }
-299
-300  /**
-301   * Write table content out to files in 
hdfs.
-302   */
-303  public static class Importer extends 
TableMapper {
-304private Map 
cfRenameMap;
-305private List 
clusterIds;
-306private Filter filter;
-307private Durability durability;
-308
-309/**
-310 * @param row  The current table row 
key.
-311 * @param value  The columns.
-312 * @param context  The current 
context.
-313 * @throws IOException When something 
is broken with the data.
-314 */
-315@Override
-316public void 
map(ImmutableBytesWritable row, Result value,
-317  Context context)
-318throws IOException {
-319  try {
-320writeResult(row, value, 
context);
-321  } catch (InterruptedException e) 
{
-322e.printStackTrace();
-323  }
-324}
-325
-326private void 
writeResult(ImmutableBytesWritable key, Result result, Context context)
-327throws IOException,

[27/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.

2018-08-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html
index b7b4236..3d1edb3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html
@@ -259,1863 +259,1867 @@
 251   * + Metadata!  + <= See note on 
BLOCK_METADATA_SPACE above.
 252   * ++
 253   * 
-254   * @see #serialize(ByteBuffer)
+254   * @see #serialize(ByteBuffer, 
boolean)
 255   */
-256  static final 
CacheableDeserializer BLOCK_DESERIALIZER =
-257  new 
CacheableDeserializer() {
-258@Override
-259public HFileBlock 
deserialize(ByteBuff buf, boolean reuse, MemoryType memType)
-260throws IOException {
-261  // The buf has the file block 
followed by block metadata.
-262  // Set limit to just before the 
BLOCK_METADATA_SPACE then rewind.
-263  buf.limit(buf.limit() - 
BLOCK_METADATA_SPACE).rewind();
-264  // Get a new buffer to pass the 
HFileBlock for it to 'own'.
-265  ByteBuff newByteBuff;
-266  if (reuse) {
-267newByteBuff = buf.slice();
-268  } else {
-269int len = buf.limit();
-270newByteBuff = new 
SingleByteBuff(ByteBuffer.allocate(len));
-271newByteBuff.put(0, buf, 
buf.position(), len);
-272  }
-273  // Read out the 
BLOCK_METADATA_SPACE content and shove into our HFileBlock.
-274  buf.position(buf.limit());
-275  buf.limit(buf.limit() + 
HFileBlock.BLOCK_METADATA_SPACE);
-276  boolean usesChecksum = buf.get() == 
(byte) 1;
-277  long offset = buf.getLong();
-278  int nextBlockOnDiskSize = 
buf.getInt();
-279  HFileBlock hFileBlock =
-280  new HFileBlock(newByteBuff, 
usesChecksum, memType, offset, nextBlockOnDiskSize, null);
-281  return hFileBlock;
-282}
-283
-284@Override
-285public int 
getDeserialiserIdentifier() {
-286  return DESERIALIZER_IDENTIFIER;
-287}
-288
-289@Override
-290public HFileBlock 
deserialize(ByteBuff b) throws IOException {
-291  // Used only in tests
-292  return deserialize(b, false, 
MemoryType.EXCLUSIVE);
-293}
-294  };
-295
-296  private static final int 
DESERIALIZER_IDENTIFIER;
-297  static {
-298DESERIALIZER_IDENTIFIER =
-299
CacheableDeserializerIdManager.registerDeserializer(BLOCK_DESERIALIZER);
-300  }
-301
-302  /**
-303   * Copy constructor. Creates a shallow 
copy of {@code that}'s buffer.
-304   */
-305  private HFileBlock(HFileBlock that) {
-306this(that, false);
-307  }
-308
-309  /**
-310   * Copy constructor. Creates a 
shallow/deep copy of {@code that}'s buffer as per the boolean
-311   * param.
-312   */
-313  private HFileBlock(HFileBlock that, 
boolean bufCopy) {
-314init(that.blockType, 
that.onDiskSizeWithoutHeader,
-315
that.uncompressedSizeWithoutHeader, that.prevBlockOffset,
-316that.offset, 
that.onDiskDataSizeWithHeader, that.nextBlockOnDiskSize, that.fileContext);
-317if (bufCopy) {
-318  this.buf = new 
SingleByteBuff(ByteBuffer.wrap(that.buf.toBytes(0, that.buf.limit(;
-319} else {
-320  this.buf = that.buf.duplicate();
-321}
-322  }
-323
-324  /**
-325   * Creates a new {@link HFile} block 
from the given fields. This constructor
-326   * is used only while writing blocks 
and caching,
-327   * and is sitting in a byte buffer and 
we want to stuff the block into cache.
-328   *
-329   * 

TODO: The caller presumes no checksumming -330 * required of this block instance since going into cache; checksum already verified on -331 * underlying block data pulled in from filesystem. Is that correct? What if cache is SSD? +256 public static final CacheableDeserializer BLOCK_DESERIALIZER = new BlockDeserializer(); +257 +258 public static final class BlockDeserializer implements CacheableDeserializer { +259private BlockDeserializer() { +260} +261 +262@Override +263public HFileBlock deserialize(ByteBuff buf, boolean reuse, MemoryType memType) +264throws IOException { +265 // The buf has the file block followed by block metadata. +266 // Set limit to just before the BLOCK_METADATA_SPACE then rewind. +267 buf.limit(buf.limit() - BLOCK_METADATA_SPACE).rewind(); +268 // Get a new buffer to pass the HFileBlock for it to 'own'. +269 ByteBuff newByteBuff; +270 if (reuse) { +271newByteBuff = buf.slice(); +272 } else { +273int len = buf.limit(); +274newByteBuff = new SingleByteBuff(ByteBuffer.allocate(len)); +275newByteBuff.put(0, buf, buf.position(), len); +276


[27/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.

2018-07-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
index 9d1748f..ac41873 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
@@ -411,7 +411,7 @@ extends AbstractStateMachineTableProcedure
-acquireLock,
 checkOnline,
 checkTableModifiable,
 getRegionDir,
 getUser, preflightChecks,
 releaseLock,
 releaseSyncLatch,
 setUser
+acquireLock,
 checkOnline,
 checkTableModifiable,
 getRegionDir,
 getUser, preflightChecks,
 releaseLock,
 releaseSyncLatch,
 setUser,
 waitInitialized<
 /li>
 
 
 
@@ -425,7 +425,7 @@ extends Procedure
-addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 doExecute,
 d
 oReleaseLock, doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics, getProcId,
 getProcIdHashCode,
 getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState, getSubmittedTime,
 getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout, haveSameParent,
 holdLock,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 <
 a 
href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#isWaiting--">isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey, setOwner,
 setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB, updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
+addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doExecute,
 doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics,
 getProcId,
 getProcIdHashCode, getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime,
 getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 holdLock, 
incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner,
 setOwner, setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout, setTimeoutFailure,
 shouldWaitClientAck,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB,
 updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html
index ad95213..941b155 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html
@@ -411,7 +411,7 @@ implements Procedure
-addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 doExecute,
 d
 oReleaseLock, doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcId, getProcIdHashCode,
 getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime,
  getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 holdLock,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner, setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB,
 updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
+addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doExecute,
 doRollback,
 elapse

[27/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.

2018-07-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
index 233dba3..91b9055 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
@@ -540,1205 +540,1204 @@
 532  
sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) -> {
 533Deque 
rLoads = oldLoads.get(Bytes.toString(regionName));
 534if (rLoads == null) {
-535  // There was nothing there
-536  rLoads = new 
ArrayDeque<>();
-537} else if (rLoads.size() >= 
numRegionLoadsToRemember) {
-538  rLoads.remove();
-539}
-540rLoads.add(new 
BalancerRegionLoad(rm));
-541
loads.put(Bytes.toString(regionName), rLoads);
-542  });
-543});
-544
-545for(CostFromRegionLoadFunction cost : 
regionLoadFunctions) {
-546  cost.setLoads(loads);
-547}
-548  }
-549
-550  protected void initCosts(Cluster 
cluster) {
-551for (CostFunction c:costFunctions) 
{
-552  c.init(cluster);
-553}
-554  }
-555
-556  protected void 
updateCostsWithAction(Cluster cluster, Action action) {
-557for (CostFunction c : costFunctions) 
{
-558  c.postAction(action);
-559}
-560  }
-561
-562  /**
-563   * Get the names of the cost 
functions
-564   */
-565  public String[] getCostFunctionNames() 
{
-566if (costFunctions == null) return 
null;
-567String[] ret = new 
String[costFunctions.length];
-568for (int i = 0; i < 
costFunctions.length; i++) {
-569  CostFunction c = 
costFunctions[i];
-570  ret[i] = 
c.getClass().getSimpleName();
-571}
-572
-573return ret;
-574  }
-575
-576  /**
-577   * This is the main cost function.  It 
will compute a cost associated with a proposed cluster
-578   * state.  All different costs will be 
combined with their multipliers to produce a double cost.
-579   *
-580   * @param cluster The state of the 
cluster
-581   * @param previousCost the previous 
cost. This is used as an early out.
-582   * @return a double of a cost 
associated with the proposed cluster state.  This cost is an
-583   * aggregate of all individual 
cost functions.
-584   */
-585  protected double computeCost(Cluster 
cluster, double previousCost) {
-586double total = 0;
-587
-588for (int i = 0; i < 
costFunctions.length; i++) {
-589  CostFunction c = 
costFunctions[i];
-590  this.tempFunctionCosts[i] = 0.0;
-591
-592  if (c.getMultiplier() <= 0) {
-593continue;
-594  }
-595
-596  Float multiplier = 
c.getMultiplier();
-597  Double cost = c.cost();
-598
-599  this.tempFunctionCosts[i] = 
multiplier*cost;
-600  total += 
this.tempFunctionCosts[i];
-601
-602  if (total > previousCost) {
-603break;
-604  }
-605}
-606
-607return total;
-608  }
-609
-610  /** Generates a candidate action to be 
applied to the cluster for cost function search */
-611  abstract static class 
CandidateGenerator {
-612abstract Cluster.Action 
generate(Cluster cluster);
-613
-614/**
-615 * From a list of regions pick a 
random one. Null can be returned which
-616 * {@link 
StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region 
move
-617 * rather than swap.
-618 *
-619 * @param clusterThe state of 
the cluster
-620 * @param server index of the 
server
-621 * @param chanceOfNoSwap Chance that 
this will decide to try a move rather
-622 *   than a 
swap.
-623 * @return a random {@link 
RegionInfo} or null if an asymmetrical move is
-624 * suggested.
-625 */
-626protected int 
pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) {
-627  // Check to see if this is just a 
move.
-628  if 
(cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat() < 
chanceOfNoSwap) {
-629// signal a move only.
-630return -1;
-631  }
-632  int rand = 
RANDOM.nextInt(cluster.regionsPerServer[server].length);
-633  return 
cluster.regionsPerServer[server][rand];
-634
-635}
-636protected int 
pickRandomServer(Cluster cluster) {
-637  if (cluster.numServers < 1) {
-638return -1;
-639  }
-640
-641  return 
RANDOM.nextInt(cluster.numServers);
-642}
-643
-644protected int pickRandomRack(Cluster 
cluster) {
-645  if (

[27/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.

2018-07-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html 
b/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
index f1708c1..8a6ff31 100644
--- a/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
+++ b/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
@@ -1,6 +1,6 @@
 http://www.w3.org/TR/html4/loose.dtd";>
 
-
+
 
 
 
@@ -19,45 +19,45 @@
 }
 //-->
 var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":9,"i15":10,"i16":9,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10};
-var tabs = 
{65535:["t0","所有方法"],1:["t1","静态方法"],2:["t2","实例方法"],8:["t4","å
…·ä½“方法"]};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
 var activeTableTab = "activeTableTab";
 
 
-您的浏览器已禁用 JavaScript。
+JavaScript is disabled on your browser.
 
 
 
 
 
-跳过导航链接
+Skip navigation links
 
 
 
-
-概览
-程序包
-ç±»
-使用
-树
-已过时
-索引
-帮助
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
 
 
 
 
-上一个类
-下一个类
+Prev Class
+Next Class
 
 
-框架
-无框架
+Frames
+No Frames
 
 
-所有类
+All Classes
 
 
 

[27/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.

2018-07-19 Thread zhangduo
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html
--
diff --git 
a/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html 
b/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html
index c0db59c..74ecf35 100644
--- a/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html
+++ b/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html
@@ -1,6 +1,6 @@
 http://www.w3.org/TR/html4/loose.dtd";>
 
-
+
 
 
 
@@ -20,38 +20,38 @@
 //-->
 
 
-JavaScript is disabled on your browser.
+您的浏览器已禁用 JavaScript。
 
 
 
 
 
-Skip navigation links
+跳过导航链接
 
 
 
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
+
+概览
+程序包
+ç±»
+使用
+树
+已过时
+索引
+帮助
 
 
 
 
-Prev Class
-Next Class
+上一个类
+下一个类
 
 
-Frames
-No Frames
+框架
+无
框架
 
 
-All Classes
+所有类
 
 
 
 
org.apache.hadoop.hbase
-

Class InvalidFamilyOperationException

+

ç±» InvalidFamilyOperationException

    -
  • https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">java.lang.Object
  • +
  • https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="java.lang中的类或接口">java.lang.Object
    • -
    • https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true"; title="class or interface in java.lang">java.lang.Throwable
    • +
    • https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true"; title="java.lang中的类或接口">java.lang.Throwable
      • -
      • https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true"; title="class or interface in java.lang">java.lang.Exception
      • +
      • https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true"; title="java.lang中的类或接口">java.lang.Exception
        • -
        • https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true"; title="class or interface in java.io">java.io.IOException
        • +
        • https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true"; title="java.io中的类或接口">java.io.IOException
          • -
          • org.apache.hadoop.hbase.HBaseIOException
          • +
          • [27/51] [partial] hbase-site git commit: Published site at 0f23784182ab88649de340d75804e0ff20dcd0fc.
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html
            index e23e8f7..cf50fff 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10};
            +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10};
             var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -186,37 +186,29 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
            long offHeapSizeDelta) 
             
             
            -void
            -blockUpdates() 
            -
            -
             https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ThreadPoolExecutor.html?is-external=true";
             title="class or interface in 
            java.util.concurrent">ThreadPoolExecutor
             getInMemoryCompactionPool() 
             
            -
            +
             long
             getMemStoreFlushSize() 
             
            -
            +
             (package private) long
             getMemStoreSize() 
             
            -
            +
             int
             getNumStores() 
             
            -
            +
             RegionInfo
             getRegionInfo() 
             
            -
            +
             WAL
             getWAL() 
             
            -
            -void
            -unblockUpdates() 
            -
             
             
             
            @@ -295,31 +287,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             Method Detail
            -
            -
            -
            -
            -
            -blockUpdates
            -public void blockUpdates()
            -
            -
            -
            -
            -
            -
            -
            -unblockUpdates
            -public void unblockUpdates()
            -
            -
             
             
             
             
             
             addMemStoreSize
            -public void addMemStoreSize(long dataSizeDelta,
            +public void addMemStoreSize(long dataSizeDelta,
             long heapSizeDelta,
             long offHeapSizeDelta)
             
            @@ -330,7 +304,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             getRegionInfo
            -public RegionInfo getRegionInfo()
            +public RegionInfo getRegionInfo()
             
             
             
            @@ -339,7 +313,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             getWAL
            -public WAL getWAL()
            +public WAL getWAL()
             
             
             
            @@ -348,7 +322,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             getInMemoryCompactionPool
            -public https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ThreadPoolExecutor.html?is-external=true";
             title="class or interface in 
            java.util.concurrent">ThreadPoolExecutor getInMemoryCompactionPool()
            +public https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ThreadPoolExecutor.html?is-external=true";
             title="class or interface in 
            java.util.concurrent">ThreadPoolExecutor getInMemoryCompactionPool()
             
             
             
            @@ -357,7 +331,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             getMemStoreFlushSize
            -public long getMemStoreFlushSize()
            +public long getMemStoreFlushSize()
             
             
             
            @@ -366,7 +340,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             getNumStores
            -public int getNumStores()
            +public int getNumStores()
             
             
             
            @@ -375,7 +349,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             getMemStoreSize
            -long getMemStoreSize()
            +long getMemStoreSize()
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html
            index 84a1668..5d9b1ff 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = {"i0":6,"i1":6,"i2":6};
            +var methods = {"i0":6,"i1":6,"i2":6,"i3":6};
             var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],4:["t3","Abstract Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
             
             
             @InterfaceAudience.Private
            -public interface ReplicationSourceService
            +public interface ReplicationSourceService
             extends ReplicationService
             A source for a replication stream has to expose this 
            service. This service allows an application
              to hook into the regionserver and watch for new transactions.
            @@ -139,12 +139,18 @@ extends 
            +ReplicationSourceManager
            +getReplicationManager()
            +Returns the replication manager
            +
            +
            +
             ReplicationPeers
             getReplicationPeers()
             Return the replication peers.
             
             
            -
            +
             SyncReplicationPeerInfoProvider
             getSyncReplicationPeerInfoProvider()
             Returns an info provider for sync replication peer.
            @@ -178,

            [27/51] [partial] hbase-site git commit: Published site at 85b41f36e01214b6485c9352875c84ebf877dab3.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html 
            b/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html
            index 2b98dff..a644a4a 100644
            --- a/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html
            +++ b/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9};
            +var methods = 
            {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9};
             var tabs = {65535:["t0","All Methods"],1:["t1","Static 
            Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
             
             
             @InterfaceAudience.Private
            -public class AccessControlUtil
            +public class AccessControlUtil
             extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object
             
             Since:
            @@ -229,13 +229,42 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +getUserPermissions(com.google.protobuf.RpcController controller,
            +  
            org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
            +  byte[] namespace,
            +  https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String userName)
            +A utility used to get permissions for selected namespace 
            based on the specified user name.
            +
            +
            +
            +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +getUserPermissions(com.google.protobuf.RpcController controller,
            +  
            org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
            +  https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String userName)
            +A utility used to get user's global permissions based on 
            the specified user name.
            +
            +
            +
            +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             getUserPermissions(com.google.protobuf.RpcController controller,
               
            org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
               TableName t)
             A utility used to get user table permissions.
             
             
            -
            +
            +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +getUserPermissions(com.google.protobuf.RpcController controller,
            +  
            org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
            +  TableName t,
            +  byte[] columnFamily,
            +  byte[] columnQualifier,
            +  https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String userName)
            +A utility used to get user table permissions based on the 
            column family, column qualifier and
            + user name.
            +
            +
            +
             static void
             grant(com.google.protobuf.RpcController controller,
              
            org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
            @@ -245,7 +274,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             A utility used to grant a user global permissions.
             
             
            -
            +
             static void
             grant(com.google.protobuf.RpcController controller,
              
            org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
            @@ -256,7 +285,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             A utility used to grant a user namespace permissions.
             
             
            -
            +
             static void
             grant(com.google.protobuf.RpcController controller,
              
            org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
            @@ -269,7 +298,20 @@ exten

            [27/51] [partial] hbase-site git commit: Published site at 6198e1fc7dfa85c3bc6b2855f9a5fb5f4b2354ff.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
            index 1410333..b56a22f 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
            @@ -124,7 +124,7 @@ var activeTableTab = "activeTableTab";
             
             @InterfaceAudience.Private
              @InterfaceStability.Evolving
            -private static class HBaseAdmin.ReplicationFuture
            +private static class HBaseAdmin.ReplicationFuture
             extends HBaseAdmin.ProcedureFutureVoid>
             
             
            @@ -241,7 +241,7 @@ extends 
             
             peerId
            -private final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String peerId
            +private final https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String peerId
             
             
             
            @@ -250,7 +250,7 @@ extends 
             
             getOperation
            -private final https://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true";
             title="class or interface in java.util.function">SupplierString> getOperation
            +private final https://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true";
             title="class or interface in java.util.function">SupplierString> getOperation
             
             
             
            @@ -267,7 +267,7 @@ extends 
             
             ReplicationFuture
            -public ReplicationFuture(HBaseAdmin admin,
            +public ReplicationFuture(HBaseAdmin admin,
              https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String peerId,
              https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true";
             title="class or interface in java.lang">Long procId,
              https://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true";
             title="class or interface in java.util.function">SupplierString> getOperation)
            @@ -287,7 +287,7 @@ extends 
             
             toString
            -public https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String toString()
            +public https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String toString()
             
             Overrides:
             https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--";
             title="class or interface in java.lang">toString in 
            class https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
            index 9732695..356 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
            @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private static class HBaseAdmin.RestoreSnapshotFuture
            +private static class HBaseAdmin.RestoreSnapshotFuture
             extends HBaseAdmin.TableFutureVoid>
             
             
            @@ -240,7 +240,7 @@ extends 
             
             RestoreSnapshotFuture
            -public RestoreSnapshotFuture(HBaseAdmin admin,
            +public RestoreSnapshotFuture(HBaseAdmin admin,
              
            org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot,
              TableName tableName,
              
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse response)
            @@ -252,7 +252,7 @@ extends 
             
             RestoreSnapshotFuture
            -public RestoreSnapshotFuture(HBaseAdmin admin,
            +public RestoreSnapshotFuture(HBaseAdmin admin,
              Tab

            [27/51] [partial] hbase-site git commit: Published site at 14087cc919da9f2e0b1a68f701f6365ad9d1d71f.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
             
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
            index b0c0fe7..08be9d0 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
            @@ -86,6 +86,10 @@
             org.apache.hadoop.hbase.master.assignment
              
             
            +
            +org.apache.hadoop.hbase.master.procedure
            + 
            +
             
             
             
            @@ -103,12 +107,27 @@
             
             
             
            -MoveRegionProcedure
            +private MoveRegionProcedure
             AssignmentManager.createMoveRegionProcedure(RegionPlan plan) 
             
            -
            -MoveRegionProcedure[]
            -AssignmentManager.createReopenProcedures(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true";
             title="class or interface in java.util">Collection regionInfo) 
            +
            +
            +
            +
            +
            +
            +Uses of MoveRegionProcedure in org.apache.hadoop.hbase.master.procedure
            +
            +Methods in org.apache.hadoop.hbase.master.procedure
             that return MoveRegionProcedure 
            +
            +Modifier and Type
            +Method and Description
            +
            +
            +
            +private MoveRegionProcedure
            +ReopenTableRegionsProcedure.createReopenProcedure(MasterProcedureEnv env,
            + HRegionLocation loc) 
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
             
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
            index f9b3413..a8b7714 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
            @@ -245,40 +245,44 @@
             RegionStates.RegionStateNode.compareTo(RegionStates.RegionStateNode other) 
             
             
            +private HRegionLocation
            +RegionStates.createRegionForReopen(RegionStates.RegionStateNode node) 
            +
            +
             protected void
             UnassignProcedure.finishTransition(MasterProcedureEnv env,
             RegionStates.RegionStateNode regionNode) 
             
            -
            +
             protected void
             AssignProcedure.finishTransition(MasterProcedureEnv env,
             RegionStates.RegionStateNode regionNode) 
             
            -
            +
             protected abstract void
             RegionTransitionProcedure.finishTransition(MasterProcedureEnv env,
             RegionStates.RegionStateNode regionNode) 
             
            -
            +
             private void
             AssignProcedure.handleFailure(MasterProcedureEnv env,
              RegionStates.RegionStateNode regionNode)
             Called when dispatch or subsequent OPEN request fail.
             
             
            -
            +
             (package private) boolean
             RegionStates.include(RegionStates.RegionStateNode node,
            boolean offline)
             Utility.
             
             
            -
            +
             private boolean
             AssignProcedure.incrementAndCheckMaxAttempts(MasterProcedureEnv env,
             RegionStates.RegionStateNode regionNode) 
             
            -
            +
             private boolean
             UnassignProcedure.isSafeToProceed(MasterProcedureEnv env,
            RegionStates.RegionStateNode regionNode,
            @@ -287,143 +291,143 @@
              unassign; e.g.
             
             
            -
            +
             protected boolean
             RegionTransitionProcedure.isServerOnline(MasterProcedureEnv env,
               RegionStates.RegionStateNode regionNode) 
             
            -
            +
             void
             AssignmentManager.markRegionAsClosed(RegionStates.RegionStateNode regionNode) 
             
            -
            +
             void
             AssignmentManager.markRegionAsClosing(RegionStates.RegionStateNode regionNode) 
             
            -
            +
             void
             AssignmentManager.markRegionAsOpened(RegionStates.RegionStateNode regionNode) 
             
            -
            +
             void
             AssignmentManager.markRegionAsOpening(RegionStates.RegionStateNode regionNode) 
             
            -
            +
             protected void
             UnassignProcedure.proceed(MasterProcedureEnv env,
            RegionStates.RegionStateNode regionNode)
             Set it up so when procedure is unsuspended, we'll move to 
            the procedure finish.
             
             
            -
            +
             protected void
             AssignmentManager.queueAssign(RegionStates.RegionStateNode regionNode)
             Add the assign operation to the assignment queue.
             
             
            -
            +
             protected boolean
             UnassignProcedure.remoteCallFailed(MasterProcedureEnv env,
             RegionStates.RegionStateNode regionNode,
             https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in 
            java.io">IOException exception) 
             
            -
            +
             protected boolean
             AssignProcedure.remoteCallFailed(MasterProcedureEnv env,
             RegionStates.RegionStateNode regionNode,
             https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in 
            java.io">IOException exception) 
             
            -
            +
             protected ab

            [27/51] [partial] hbase-site git commit: Published site at 72784c2d836a4b977667449d3adec5e8d15453f5.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
            index b6e7636..592c2cc 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
            @@ -356,3901 +356,3924 @@
             348  public Future 
            modifyTableAsync(TableDescriptor td) throws IOException {
             349ModifyTableResponse response = 
            executeCallable(
             350  new 
            MasterCallable(getConnection(), 
            getRpcControllerFactory()) {
            -351@Override
            -352protected ModifyTableResponse 
            rpcCall() throws Exception {
            -353  
            setPriority(td.getTableName());
            -354  ModifyTableRequest request = 
            RequestConverter.buildModifyTableRequest(
            -355td.getTableName(), td, 
            ng.getNonceGroup(), ng.newNonce());
            -356  return 
            master.modifyTable(getRpcController(), request);
            -357}
            -358  });
            -359return new ModifyTableFuture(this, 
            td.getTableName(), response);
            -360  }
            -361
            -362  @Override
            -363  public List 
            listTableDescriptorsByNamespace(byte[] name) throws IOException {
            -364return executeCallable(new 
            MasterCallable>(getConnection(),
            -365getRpcControllerFactory()) {
            -366  @Override
            -367  protected 
            List rpcCall() throws Exception {
            -368return 
            master.listTableDescriptorsByNamespace(getRpcController(),
            -369
            ListTableDescriptorsByNamespaceRequest.newBuilder()
            -370  
            .setNamespaceName(Bytes.toString(name)).build())
            -371.getTableSchemaList()
            -372.stream()
            -373
            .map(ProtobufUtil::toTableDescriptor)
            -374
            .collect(Collectors.toList());
            -375  }
            -376});
            -377  }
            -378
            -379  @Override
            -380  public List 
            listTableDescriptors(List tableNames) throws IOException {
            -381return executeCallable(new 
            MasterCallable>(getConnection(),
            -382getRpcControllerFactory()) {
            -383  @Override
            -384  protected 
            List rpcCall() throws Exception {
            -385GetTableDescriptorsRequest req 
            =
            -386
            RequestConverter.buildGetTableDescriptorsRequest(tableNames);
            -387  return 
            ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(),
            -388  req));
            -389  }
            -390});
            -391  }
            -392
            -393  @Override
            -394  public List 
            getRegions(final ServerName sn) throws IOException {
            -395AdminService.BlockingInterface admin 
            = this.connection.getAdmin(sn);
            -396// TODO: There is no timeout on this 
            controller. Set one!
            -397HBaseRpcController controller = 
            rpcControllerFactory.newController();
            -398return 
            ProtobufUtil.getOnlineRegions(controller, admin);
            -399  }
            -400
            -401  @Override
            -402  public List 
            getRegions(TableName tableName) throws IOException {
            -403if 
            (TableName.isMetaTableName(tableName)) {
            -404  return 
            Arrays.asList(RegionInfoBuilder.FIRST_META_REGIONINFO);
            -405} else {
            -406  return 
            MetaTableAccessor.getTableRegions(connection, tableName, true);
            -407}
            -408  }
            -409
            -410  private static class 
            AbortProcedureFuture extends ProcedureFuture {
            -411private boolean isAbortInProgress;
            -412
            -413public AbortProcedureFuture(
            -414final HBaseAdmin admin,
            -415final Long procId,
            -416final Boolean abortProcResponse) 
            {
            -417  super(admin, procId);
            -418  this.isAbortInProgress = 
            abortProcResponse;
            -419}
            -420
            -421@Override
            -422public Boolean get(long timeout, 
            TimeUnit unit)
            -423throws InterruptedException, 
            ExecutionException, TimeoutException {
            -424  if (!this.isAbortInProgress) {
            -425return false;
            -426  }
            -427  super.get(timeout, unit);
            -428  return true;
            -429}
            -430  }
            -431
            -432  /** @return Connection used by this 
            object. */
            -433  @Override
            -434  public Connection getConnection() {
            -435return connection;
            -436  }
            -437
            -438  @Override
            -439  public boolean tableExists(final 
            TableName tableName) throws IOException {
            -440return executeCallable(new 
            RpcRetryingCallable() {
            -441  @Override
            -442  protected Boolean rpcCall(int 
            callTimeout) throws Exception {
            -443return 
            MetaTableAccessor.tableExists(connection, tableName);
            -444  }
            -445});
            -446  }
            -447
            -448  @Override
            -449  public HTableDescriptor[] listTables() 
            throws IOException {
            -450return listTables((Pattern)null, 
            false);
            -451  }
            -452
            -453  @Override
            -454  public HTableDescriptor[] 
            listTables(Pattern pa

            [27/51] [partial] hbase-site git commit: Published site at 9101fc246f86445006bfbcdfda5cc495016dc280.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html 
            b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
            index a9d814d..e65cf97 100644
            --- a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
            +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10};
            +var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10};
             var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -288,13 +288,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             private https://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true";
             title="class or interface in java.lang">ThreadGroup
             threadGroup
            -Created in the start(int,
             boolean) method.
            +Created in the #start(int, boolean) 
            method.
             
             
             
             private TimeoutExecutorThread
             timeoutExecutor
            -Created in the start(int,
             boolean) method.
            +Created in the #start(int, boolean) 
            method.
             
             
             
            @@ -308,7 +308,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CopyOnWriteArrayList.html?is-external=true";
             title="class or interface in 
            java.util.concurrent">CopyOnWriteArrayList
             workerThreads
            -Created in the start(int,
             boolean) method.
            +Created in the #start(int, boolean) 
            method.
             
             
             
            @@ -490,107 +490,114 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
               https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true";
             title="class or interface in 
            java.lang">InterruptedException e) 
             
             
            +void
            +init(int numThreads,
            +boolean abortOnCorruption)
            +Initialize the procedure executor, but do not start 
            workers.
            +
            +
            +
             private Procedure[]
             initializeChildren(RootProcedureState procStack,
               Procedure procedure,
               Procedure[] subprocs) 
             
            -
            +
             boolean
             isFinished(long procId)
             Return true if the procedure is finished.
             
             
            -
            +
             boolean
             isProcedureOwner(long procId,
             User user)
             Check if the user is this procedure's owner
             
             
            -
            +
             boolean
             isRunning() 
             
            -
            +
             boolean
             isStarted(long procId)
             Return true if the procedure is started.
             
             
            -
            +
             void
             join() 
             
            -
            +
             private void
             load(boolean abortOnCorruption) 
             
            -
            +
             private void
             loadProcedures(ProcedureStore.ProcedureIterator procIter,
               boolean abortOnCorruption) 
             
            -
            +
             private long
             nextProcId() 
             
            -
            +
             private Procedure
             prepareProcedure(Procedure proc) 
             
            -
            +
             private void
             procedureFinished(Procedure proc) 
             
            -
            +
             private long
             pushProcedure(Procedure proc) 
             
            -
            +
             void
             refreshConfiguration(org.apache.hadoop.conf.Configuration conf) 
             
            -
            +
             void
             registerListener(ProcedureExecutor.ProcedureExecutorListener listener) 
             
            -
            +
             long
             registerNonce(NonceKey nonceKey)
             Register a nonce for a procedure that is going to be 
            submitted.
             
             
            -
            +
             private void
             releaseLock(Procedure proc,
            boolean force) 
             
            -
            +
             boolean
             removeChore(ProcedureInMemoryChore chore)
             Remove a chore procedure from the executor
             
             
            -
            +
             void
             removeResult(long procId)
             Mark the specified completed procedure, as ready to 
            remove.
             
             
            -
            +
             private void
             sendProcedureAddedNotification(long procId) 
             
            -
            +
             private void
             sendProcedureFinishedNotification(long procId) 
             
            -
            +
             private void
             sendProcedureLoadedNotification(long procId) 
             
            -
            +
             void
             setFailureResultForNonce(NonceKey nonceKey,
             https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";

            [27/51] [partial] hbase-site git commit: Published site at 0b28155d274910b4e667b949d51f78809a1eff0b.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
            index f236300..513d2ad 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
            @@ -26,1048 +26,1115 @@
             018
             019package 
            org.apache.hadoop.hbase.backup.impl;
             020
            -021import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
            -022import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
            -023import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
            -024import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
            -025import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
            -026import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
            -027import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
            -028import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
            -029import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
            -030import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
            -031import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
            -032import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
            -033import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
            -034import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
            -035import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
            -036import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
            -037import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
            -038import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
            -039
            -040import java.io.IOException;
            -041import java.net.URI;
            -042import java.util.List;
            +021import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BACKUP_LIST_DESC;
            +022import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
            +023import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
            +024import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
            +025import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
            +026import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP;
            +027import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP_DESC;
            +028import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_LIST;
            +029import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
            +030import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
            +031import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
            +032import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
            +033import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
            +034import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
            +035import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
            +036import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
            +037import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
            +038import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
            +039import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
            +040import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
            +041import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
            +042import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
             043
            -044import 
            org.apache.commons.lang3.StringUtils;
            -045import 
            org.apache.hadoop.conf.Configuration;
            -046import 
            org.apache.hadoop.conf.Configured;
            -047import org.apache.hadoop.fs.FileSystem;
            -048import org.apache.hadoop.fs.Path;
            -049import 
            org.apache.hadoop.hbase.HBaseConfiguration;
            -050import 
            org.apache.hadoop.hbase.TableName;
            -051import 
            org.apache.hadoop.hbase.backup.BackupAdmin;
            -052import 
            org.apache.hadoop.hbase.backup.BackupInfo;
            -053import 
            org.apache.h

            [27/51] [partial] hbase-site git commit: Published site at 7d3750bd9fc9747623549c242cc4171e224b3eaf.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            index 275ca84..b9a30c4 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            @@ -1499,782 +1499,786 @@
             1491  final RpcController controller,
             1492  final IsInMaintenanceModeRequest 
            request) throws ServiceException {
             1493IsInMaintenanceModeResponse.Builder 
            response = IsInMaintenanceModeResponse.newBuilder();
            -1494
            response.setInMaintenanceMode(master.isInMaintenanceMode());
            -1495return response.build();
            -1496  }
            -1497
            -1498  @Override
            -1499  public UnassignRegionResponse 
            unassignRegion(RpcController controller,
            -1500  UnassignRegionRequest req) throws 
            ServiceException {
            -1501try {
            -1502  final byte [] regionName = 
            req.getRegion().getValue().toByteArray();
            -1503  RegionSpecifierType type = 
            req.getRegion().getType();
            -1504  final boolean force = 
            req.getForce();
            -1505  UnassignRegionResponse urr = 
            UnassignRegionResponse.newBuilder().build();
            -1506
            -1507  master.checkInitialized();
            -1508  if (type != 
            RegionSpecifierType.REGION_NAME) {
            -1509LOG.warn("unassignRegion 
            specifier type: expected: " + RegionSpecifierType.REGION_NAME
            -1510  + " actual: " + type);
            -1511  }
            -1512  Pair 
            pair =
            -1513
            MetaTableAccessor.getRegion(master.getConnection(), regionName);
            -1514  if 
            (Bytes.equals(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(),regionName))
             {
            -1515pair = new 
            Pair<>(RegionInfoBuilder.FIRST_META_REGIONINFO,
            -1516
            master.getMetaTableLocator().getMetaRegionLocation(master.getZooKeeper()));
            -1517  }
            -1518  if (pair == null) {
            -1519throw new 
            UnknownRegionException(Bytes.toString(regionName));
            -1520  }
            -1521
            -1522  RegionInfo hri = 
            pair.getFirst();
            -1523  if (master.cpHost != null) {
            -1524master.cpHost.preUnassign(hri, 
            force);
            -1525  }
            -1526  
            LOG.debug(master.getClientIdAuditPrefix() + " unassign " + 
            hri.getRegionNameAsString()
            -1527  + " in current location if it 
            is online and reassign.force=" + force);
            -1528  
            master.getAssignmentManager().unassign(hri);
            -1529  if (master.cpHost != null) {
            -1530master.cpHost.postUnassign(hri, 
            force);
            -1531  }
            -1532
            -1533  return urr;
            -1534} catch (IOException ioe) {
            -1535  throw new ServiceException(ioe);
            -1536}
            -1537  }
            -1538
            -1539  @Override
            -1540  public 
            ReportRegionStateTransitionResponse reportRegionStateTransition(RpcController 
            c,
            -1541  ReportRegionStateTransitionRequest 
            req) throws ServiceException {
            -1542try {
            -1543  master.checkServiceStarted();
            -1544  return 
            master.getAssignmentManager().reportRegionStateTransition(req);
            -1545} catch (IOException ioe) {
            -1546  throw new ServiceException(ioe);
            -1547}
            -1548  }
            -1549
            -1550  @Override
            -1551  public SetQuotaResponse 
            setQuota(RpcController c, SetQuotaRequest req)
            -1552  throws ServiceException {
            -1553try {
            -1554  master.checkInitialized();
            -1555  return 
            master.getMasterQuotaManager().setQuota(req);
            -1556} catch (Exception e) {
            -1557  throw new ServiceException(e);
            -1558}
            -1559  }
            -1560
            -1561  @Override
            -1562  public 
            MajorCompactionTimestampResponse getLastMajorCompactionTimestamp(RpcController 
            controller,
            -1563  MajorCompactionTimestampRequest 
            request) throws ServiceException {
            -1564
            MajorCompactionTimestampResponse.Builder response =
            -1565
            MajorCompactionTimestampResponse.newBuilder();
            -1566try {
            -1567  master.checkInitialized();
            -1568  
            response.setCompactionTimestamp(master.getLastMajorCompactionTimestamp(ProtobufUtil
            -1569  
            .toTableName(request.getTableName(;
            -1570} catch (IOException e) {
            -1571  throw new ServiceException(e);
            -1572}
            -1573return response.build();
            -1574  }
            -1575
            -1576  @Override
            -1577  public 
            MajorCompactionTimestampResponse getLastMajorCompactionTimestampForRegion(
            -1578  RpcController controller, 
            MajorCompactionTimestampForRegionRequest request)
            -1579  throws ServiceException {
            -1580
            MajorCompactionTimestampResponse.Builder response =
            -1581
            MajorCompactionTimestampResponse.newBuilder();
            -1582try {
            -1583  master.checkInitialized();
            -1584  
            response.setCompactionTimestamp(master.getLastMajorCompactionTimestampForRegion(request
            -1585  
            .getRegion().getValue().toByteArray()));
            -1586} catch (IOException e) {
            -1587  throw new ServiceE

            [27/51] [partial] hbase-site git commit: Published site at 997747076d8ec0b4346d7cb99c4b0667a7c14905.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
            index 3da432b..d30fa8f 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
            @@ -928,7690 +928,7698 @@
             920  Collection stores = 
            this.stores.values();
             921  try {
             922// update the stores that we are 
            replaying
            -923
            stores.forEach(HStore::startReplayingFromWAL);
            -924// Recover any edits if 
            available.
            -925maxSeqId = Math.max(maxSeqId,
            -926  
            replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
            status));
            -927// Make sure mvcc is up to max.
            -928this.mvcc.advanceTo(maxSeqId);
            -929  } finally {
            -930// update the stores that we are 
            done replaying
            -931
            stores.forEach(HStore::stopReplayingFromWAL);
            -932  }
            -933}
            -934this.lastReplayedOpenRegionSeqId = 
            maxSeqId;
            -935
            -936
            this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
            -937this.writestate.flushRequested = 
            false;
            -938this.writestate.compacting.set(0);
            -939
            -940if (this.writestate.writesEnabled) 
            {
            -941  // Remove temporary data left over 
            from old regions
            -942  status.setStatus("Cleaning up 
            temporary data from old regions");
            -943  fs.cleanupTempDir();
            -944}
            -945
            -946if (this.writestate.writesEnabled) 
            {
            -947  status.setStatus("Cleaning up 
            detritus from prior splits");
            -948  // Get rid of any splits or merges 
            that were lost in-progress.  Clean out
            -949  // these directories here on open.  
            We may be opening a region that was
            -950  // being split but we crashed in 
            the middle of it all.
            -951  fs.cleanupAnySplitDetritus();
            -952  fs.cleanupMergesDir();
            -953}
            -954
            -955// Initialize split policy
            -956this.splitPolicy = 
            RegionSplitPolicy.create(this, conf);
            -957
            -958// Initialize flush policy
            -959this.flushPolicy = 
            FlushPolicyFactory.create(this, conf);
            -960
            -961long lastFlushTime = 
            EnvironmentEdgeManager.currentTime();
            -962for (HStore store: stores.values()) 
            {
            -963  
            this.lastStoreFlushTimeMap.put(store, lastFlushTime);
            -964}
            -965
            -966// Use maximum of log sequenceid or 
            that which was found in stores
            -967// (particularly if no recovered 
            edits, seqid will be -1).
            -968long maxSeqIdFromFile =
            -969  
            WALSplitter.getMaxRegionSequenceId(fs.getFileSystem(), fs.getRegionDir());
            -970long nextSeqId = Math.max(maxSeqId, 
            maxSeqIdFromFile) + 1;
            -971if (writestate.writesEnabled) {
            -972  
            WALSplitter.writeRegionSequenceIdFile(fs.getFileSystem(), fs.getRegionDir(), 
            nextSeqId - 1);
            -973}
            -974
            -975LOG.info("Opened {}; next 
            sequenceid={}", this.getRegionInfo().getShortNameToLog(), nextSeqId);
            -976
            -977// A region can be reopened if failed 
            a split; reset flags
            -978this.closing.set(false);
            -979this.closed.set(false);
            -980
            -981if (coprocessorHost != null) {
            -982  status.setStatus("Running 
            coprocessor post-open hooks");
            -983  coprocessorHost.postOpen();
            -984}
            +923LOG.debug("replaying wal for " + 
            this.getRegionInfo().getEncodedName());
            +924
            stores.forEach(HStore::startReplayingFromWAL);
            +925// Recover any edits if 
            available.
            +926maxSeqId = Math.max(maxSeqId,
            +927  
            replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
            status));
            +928// Make sure mvcc is up to max.
            +929this.mvcc.advanceTo(maxSeqId);
            +930  } finally {
            +931LOG.debug("stopping wal replay 
            for " + this.getRegionInfo().getEncodedName());
            +932// update the stores that we are 
            done replaying
            +933
            stores.forEach(HStore::stopReplayingFromWAL);
            +934  }
            +935}
            +936this.lastReplayedOpenRegionSeqId = 
            maxSeqId;
            +937
            +938
            this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
            +939this.writestate.flushRequested = 
            false;
            +940this.writestate.compacting.set(0);
            +941
            +942if (this.writestate.writesEnabled) 
            {
            +943  LOG.debug("Cleaning up temporary 
            data for " + this.getRegionInfo().getEncodedName());
            +944  // Remove temporary data left over 
            from old regions
            +945  status.setStatus("Cleaning up 
            temporary data from old regions");
            +946  fs.cleanupTempDir();
            +947}
            +948
            +949if (this.writestate.writesEnabled) 
            {
            +950  status.setStatus("Cleaning up 
            detritus from prior splits");
            +951  // Get rid of any splits or merges 
            that were lost in-progress.  Clean out
            +952  // these directories here on open.  
            

            [27/51] [partial] hbase-site git commit: Published site at f3d1c021de2264301f68eadb9ef126ff83d7ef53.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
            index 7d27402..3da432b 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
            @@ -2560,8 +2560,8 @@
             2552}
             2553MemStoreSize mss = 
            this.memStoreSizing.getMemStoreSize();
             2554LOG.info("Flushing " + + 
            storesToFlush.size() + "/" + stores.size() + " column families," +
            -2555" memstore data size=" + 
            StringUtils.byteDesc(mss.getDataSize()) +
            -2556" memstore heap size=" + 
            StringUtils.byteDesc(mss.getHeapSize()) +
            +2555" dataSize=" + 
            StringUtils.byteDesc(mss.getDataSize()) +
            +2556" heapSize=" + 
            StringUtils.byteDesc(mss.getHeapSize()) +
             2557((perCfExtras != null && 
            perCfExtras.length() > 0)? perCfExtras.toString(): "") +
             2558((wal != null) ? "" : "; WAL is 
            null, using passed sequenceid=" + sequenceId));
             2559  }
            @@ -2761,10 +2761,10 @@
             2753long time = 
            EnvironmentEdgeManager.currentTime() - startTime;
             2754MemStoreSize mss = 
            prepareResult.totalFlushableSize.getMemStoreSize();
             2755long memstoresize = 
            this.memStoreSizing.getMemStoreSize().getDataSize();
            -2756String msg = "Finished memstore 
            flush;"
            -2757+ " data size ~" + 
            StringUtils.byteDesc(mss.getDataSize()) + "/" + mss.getDataSize()
            -2758+ ", heap size ~" + 
            StringUtils.byteDesc(mss.getHeapSize()) + "/" + mss.getHeapSize()
            -2759+ ", currentsize=" + 
            StringUtils.byteDesc(memstoresize) + "/" + memstoresize
            +2756String msg = "Finished flush of"
            +2757+ " dataSize ~" + 
            StringUtils.byteDesc(mss.getDataSize()) + "/" + mss.getDataSize()
            +2758+ ", heapSize ~" + 
            StringUtils.byteDesc(mss.getHeapSize()) + "/" + mss.getHeapSize()
            +2759+ ", currentSize=" + 
            StringUtils.byteDesc(memstoresize) + "/" + memstoresize
             2760+ " for " + 
            this.getRegionInfo().getEncodedName() + " in " + time + "ms, sequenceid="
             2761+ flushOpSeqId +  ", compaction 
            requested=" + compactionRequested
             2762+ ((wal == null) ? "; wal=null" 
            : "");
            @@ -4244,7 +4244,7 @@
             4236   * @param cellItr
             4237   * @param now
             4238   */
            -4239  public void updateCellTimestamps(final 
            Iterable> cellItr, final byte[] now)
            +4239  private static void 
            updateCellTimestamps(final Iterable> cellItr, final 
            byte[] now)
             4240  throws IOException {
             4241for (List cells: 
            cellItr) {
             4242  if (cells == null) continue;
            @@ -4299,12 +4299,12 @@
             4291  requestFlush();
             4292  // Don't print current limit 
            because it will vary too much. The message is used as a key
             4293  // over in 
            RetriesExhaustedWithDetailsException processing.
            -4294  throw new 
            RegionTooBusyException("Over memstore limit; regionName=" +
            -4295  (this.getRegionInfo() == null? 
            "unknown": this.getRegionInfo().getEncodedName()) +
            -4296  ", server=" + 
            (this.getRegionServerServices() == null ? "unknown":
            -4297  
            this.getRegionServerServices().getServerName()) +
            -4298  ", blockingMemStoreSize=" +
            -4299  
            org.apache.hadoop.hbase.procedure2.util.StringUtils.humanSize(blockingMemStoreSize));
            +4294  throw new 
            RegionTooBusyException("Over memstore limit=" +
            +4295
            org.apache.hadoop.hbase.procedure2.util.StringUtils.humanSize(this.blockingMemStoreSize)
             +
            +4296", regionName=" +
            +4297  (this.getRegionInfo() == null? 
            "unknown": this.getRegionInfo().getEncodedName()) +
            +4298  ", server=" + 
            (this.getRegionServerServices() == null? "unknown":
            +4299  
            this.getRegionServerServices().getServerName()));
             4300}
             4301  }
             4302
            @@ -8610,7 +8610,8 @@
             8602  public void 
            requestFlush(FlushLifeCycleTracker tracker) throws IOException {
             8603requestFlush0(tracker);
             8604  }
            -8605}
            +8605
            +8606}
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
            index 7d27402..3da432b 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
            @@ -2560,8 +2560,8 @@
             2552}
             2553MemStoreSize mss = 
            this.memStoreSizing.getMemStoreSize(

            [27/51] [partial] hbase-site git commit: Published site at cf529f18a9959589fa635f78df4840472526ea2c.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
            --
            diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
            index a97dfdc..2b1b6c6 100644
            --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
            +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
            @@ -2370,1287 +2370,1292 @@
             2362  }
             2363
             2364  @Override
            -2365  public long modifyTable(final 
            TableName tableName, final TableDescriptor descriptor,
            +2365  public long modifyTable(final 
            TableName tableName, final TableDescriptor newDescriptor,
             2366  final long nonceGroup, final long 
            nonce) throws IOException {
             2367checkInitialized();
            -2368
            sanityCheckTableDescriptor(descriptor);
            +2368
            sanityCheckTableDescriptor(newDescriptor);
             2369
             2370return 
            MasterProcedureUtil.submitProcedure(
             2371new 
            MasterProcedureUtil.NonceProcedureRunnable(this, nonceGroup, nonce) {
             2372  @Override
             2373  protected void run() throws 
            IOException {
            -2374
            getMaster().getMasterCoprocessorHost().preModifyTable(tableName, descriptor);
            -2375
            -2376
            LOG.info(getClientIdAuditPrefix() + " modify " + tableName);
            +2374TableDescriptor oldDescriptor = 
            getMaster().getTableDescriptors().get(tableName);
            +2375
            getMaster().getMasterCoprocessorHost()
            +2376  .preModifyTable(tableName, 
            oldDescriptor, newDescriptor);
             2377
            -2378// Execute the operation 
            synchronously - wait for the operation completes before continuing.
            -2379//
            -2380// We need to wait for the 
            procedure to potentially fail due to "prepare" sanity
            -2381// checks. This will block only 
            the beginning of the procedure. See HBASE-19953.
            -2382ProcedurePrepareLatch latch = 
            ProcedurePrepareLatch.createBlockingLatch();
            -2383submitProcedure(new 
            ModifyTableProcedure(procedureExecutor.getEnvironment(),
            -2384descriptor, latch));
            -2385latch.await();
            -2386
            -2387
            getMaster().getMasterCoprocessorHost().postModifyTable(tableName, 
            descriptor);
            -2388  }
            -2389
            -2390  @Override
            -2391  protected String getDescription() 
            {
            -2392return "ModifyTableProcedure";
            -2393  }
            -2394});
            -2395  }
            -2396
            -2397  public long restoreSnapshot(final 
            SnapshotDescription snapshotDesc,
            -2398  final long nonceGroup, final long 
            nonce, final boolean restoreAcl) throws IOException {
            -2399checkInitialized();
            -2400
            getSnapshotManager().checkSnapshotSupport();
            -2401
            -2402// Ensure namespace exists. Will 
            throw exception if non-known NS.
            -2403final TableName dstTable = 
            TableName.valueOf(snapshotDesc.getTable());
            -2404
            getClusterSchema().getNamespace(dstTable.getNamespaceAsString());
            -2405
            -2406return 
            MasterProcedureUtil.submitProcedure(
            -2407new 
            MasterProcedureUtil.NonceProcedureRunnable(this, nonceGroup, nonce) {
            -2408  @Override
            -2409  protected void run() throws 
            IOException {
            -2410  setProcId(
            -2411
            getSnapshotManager().restoreOrCloneSnapshot(snapshotDesc, getNonceKey(), 
            restoreAcl));
            -2412  }
            -2413
            -2414  @Override
            -2415  protected String getDescription() 
            {
            -2416return 
            "RestoreSnapshotProcedure";
            -2417  }
            -2418});
            -2419  }
            -2420
            -2421  private void checkTableExists(final 
            TableName tableName)
            -2422  throws IOException, 
            TableNotFoundException {
            -2423if 
            (!MetaTableAccessor.tableExists(getConnection(), tableName)) {
            -2424  throw new 
            TableNotFoundException(tableName);
            -2425}
            -2426  }
            -2427
            -2428  @Override
            -2429  public void checkTableModifiable(final 
            TableName tableName)
            -2430  throws IOException, 
            TableNotFoundException, TableNotDisabledException {
            -2431if (isCatalogTable(tableName)) {
            -2432  throw new IOException("Can't 
            modify catalog tables");
            -2433}
            -2434checkTableExists(tableName);
            -2435TableState ts = 
            getTableStateManager().getTableState(tableName);
            -2436if (!ts.isDisabled()) {
            -2437  throw new 
            TableNotDisabledException("Not DISABLED; " + ts);
            -2438}
            -2439  }
            -2440
            -2441  public ClusterMetrics 
            getClusterMetricsWithoutCoprocessor() throws InterruptedIOException {
            -2442return 
            getClusterMetricsWithoutCoprocessor(EnumSet.allOf(Option.class));
            -2443  }
            -2444
            -2445  public ClusterMetrics 
            getClusterMetricsWithoutCoprocessor(EnumSet

            [27/51] [partial] hbase-site git commit: Published site at 021f66d11d2cbb7308308093e29e69d6e7661ee9.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
            index 4a879bb..7d27402 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
            @@ -300,7 +300,7 @@
             292  private Map coprocessorServiceHandlers = 
            Maps.newHashMap();
             293
             294  // Track data size in all memstores
            -295  private final MemStoreSizing 
            memStoreSize = new MemStoreSizing();
            +295  private final MemStoreSizing 
            memStoreSizing = new ThreadSafeMemStoreSizing();
             296  private final RegionServicesForStores 
            regionServicesForStores = new RegionServicesForStores(this);
             297
             298  // Debug possible data loss due to WAL 
            off
            @@ -1218,7389 +1218,7399 @@
             1210   * Increase the size of mem store in 
            this region and the size of global mem
             1211   * store
             1212   */
            -1213  public void 
            incMemStoreSize(MemStoreSize memStoreSize) {
            -1214if (this.rsAccounting != null) {
            -1215  
            rsAccounting.incGlobalMemStoreSize(memStoreSize);
            -1216}
            -1217long dataSize;
            -1218synchronized (this.memStoreSize) {
            -1219  
            this.memStoreSize.incMemStoreSize(memStoreSize);
            -1220  dataSize = 
            this.memStoreSize.getDataSize();
            -1221}
            -1222
            checkNegativeMemStoreDataSize(dataSize, memStoreSize.getDataSize());
            -1223  }
            -1224
            -1225  public void 
            decrMemStoreSize(MemStoreSize memStoreSize) {
            -1226if (this.rsAccounting != null) {
            -1227  
            rsAccounting.decGlobalMemStoreSize(memStoreSize);
            -1228}
            -1229long size;
            -1230synchronized (this.memStoreSize) {
            -1231  
            this.memStoreSize.decMemStoreSize(memStoreSize);
            -1232  size = 
            this.memStoreSize.getDataSize();
            +1213  void incMemStoreSize(MemStoreSize mss) 
            {
            +1214incMemStoreSize(mss.getDataSize(), 
            mss.getHeapSize(), mss.getOffHeapSize());
            +1215  }
            +1216
            +1217  void incMemStoreSize(long 
            dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) {
            +1218if (this.rsAccounting != null) {
            +1219  
            rsAccounting.incGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, 
            offHeapSizeDelta);
            +1220}
            +1221long dataSize =
            +1222
            this.memStoreSizing.incMemStoreSize(dataSizeDelta, heapSizeDelta, 
            offHeapSizeDelta);
            +1223
            checkNegativeMemStoreDataSize(dataSize, dataSizeDelta);
            +1224  }
            +1225
            +1226  void decrMemStoreSize(MemStoreSize 
            mss) {
            +1227decrMemStoreSize(mss.getDataSize(), 
            mss.getHeapSize(), mss.getOffHeapSize());
            +1228  }
            +1229
            +1230  void decrMemStoreSize(long 
            dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) {
            +1231if (this.rsAccounting != null) {
            +1232  
            rsAccounting.decGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, 
            offHeapSizeDelta);
             1233}
            -1234checkNegativeMemStoreDataSize(size, 
            -memStoreSize.getDataSize());
            -1235  }
            -1236
            -1237  private void 
            checkNegativeMemStoreDataSize(long memStoreDataSize, long delta) {
            -1238// This is extremely bad if we make 
            memStoreSize negative. Log as much info on the offending
            -1239// caller as possible. (memStoreSize 
            might be a negative value already -- freeing memory)
            -1240if (memStoreDataSize < 0) {
            -1241  LOG.error("Asked to modify this 
            region's (" + this.toString()
            -1242  + ") memStoreSize to a 
            negative value which is incorrect. Current memStoreSize="
            -1243  + (memStoreDataSize - delta) + 
            ", delta=" + delta, new Exception());
            -1244}
            -1245  }
            -1246
            -1247  @Override
            -1248  public RegionInfo getRegionInfo() {
            -1249return this.fs.getRegionInfo();
            -1250  }
            -1251
            -1252  /**
            -1253   * @return Instance of {@link 
            RegionServerServices} used by this HRegion.
            -1254   * Can be null.
            -1255   */
            -1256  RegionServerServices 
            getRegionServerServices() {
            -1257return this.rsServices;
            -1258  }
            -1259
            -1260  @Override
            -1261  public long getReadRequestsCount() {
            -1262return readRequestsCount.sum();
            -1263  }
            -1264
            -1265  @Override
            -1266  public long 
            getFilteredReadRequestsCount() {
            -1267return 
            filteredReadRequestsCount.sum();
            -1268  }
            -1269
            -1270  @Override
            -1271  public long getWriteRequestsCount() 
            {
            -1272return writeRequestsCount.sum();
            -1273  }
            -1274
            -1275  @Override
            -1276  public long getMemStoreDataSize() {
            -1277return memStoreSize.getDataSize();
            -1278  }
            -1279
            -1280  @Override
            -1281  public long getMemStoreHeapSize() {
            -1282return memStoreSize.getHeapSize();
            -1283  }
            -1284
            -1285  @Override
            -1286  public long getMemStoreOffHeapSize() 
            {
            -1287return 
            memStoreSize.getOffHeapSize();
            -1288  }
            -1289
            -1290  /** @return store services for this 
            region, to access services required by store level

            [27/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
            --
            diff --git 
            a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
             
            b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
            index 2510283..418c60c 100644
            --- 
            a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
            +++ 
            b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
            @@ -77,77 +77,77 @@
             069import 
            org.apache.hadoop.hbase.client.RowMutations;
             070import 
            org.apache.hadoop.hbase.client.Scan;
             071import 
            org.apache.hadoop.hbase.client.Table;
            -072import 
            org.apache.hadoop.hbase.filter.BinaryComparator;
            -073import 
            org.apache.hadoop.hbase.filter.Filter;
            -074import 
            org.apache.hadoop.hbase.filter.FilterAllFilter;
            -075import 
            org.apache.hadoop.hbase.filter.FilterList;
            -076import 
            org.apache.hadoop.hbase.filter.PageFilter;
            -077import 
            org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
            -078import 
            org.apache.hadoop.hbase.filter.WhileMatchFilter;
            -079import 
            org.apache.hadoop.hbase.io.compress.Compression;
            -080import 
            org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
            -081import 
            org.apache.hadoop.hbase.io.hfile.RandomDistribution;
            -082import 
            org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
            -083import 
            org.apache.hadoop.hbase.regionserver.BloomType;
            -084import 
            org.apache.hadoop.hbase.regionserver.CompactingMemStore;
            -085import 
            org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration;
            -086import 
            org.apache.hadoop.hbase.trace.SpanReceiverHost;
            -087import 
            org.apache.hadoop.hbase.trace.TraceUtil;
            -088import 
            org.apache.hadoop.hbase.util.ByteArrayHashKey;
            -089import 
            org.apache.hadoop.hbase.util.Bytes;
            -090import 
            org.apache.hadoop.hbase.util.Hash;
            -091import 
            org.apache.hadoop.hbase.util.MurmurHash;
            -092import 
            org.apache.hadoop.hbase.util.Pair;
            -093import 
            org.apache.hadoop.hbase.util.YammerHistogramUtils;
            -094import 
            org.apache.hadoop.io.LongWritable;
            -095import org.apache.hadoop.io.Text;
            -096import org.apache.hadoop.mapreduce.Job;
            -097import 
            org.apache.hadoop.mapreduce.Mapper;
            -098import 
            org.apache.hadoop.mapreduce.lib.input.NLineInputFormat;
            -099import 
            org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
            -100import 
            org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer;
            -101import org.apache.hadoop.util.Tool;
            -102import 
            org.apache.hadoop.util.ToolRunner;
            -103import 
            org.apache.htrace.core.ProbabilitySampler;
            -104import org.apache.htrace.core.Sampler;
            -105import 
            org.apache.htrace.core.TraceScope;
            -106import 
            org.apache.yetus.audience.InterfaceAudience;
            -107import org.slf4j.Logger;
            -108import org.slf4j.LoggerFactory;
            -109import 
            org.apache.hbase.thirdparty.com.google.common.base.MoreObjects;
            -110import 
            org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
            -111
            -112/**
            -113 * Script used evaluating HBase 
            performance and scalability.  Runs a HBase
            -114 * client that steps through one of a set 
            of hardcoded tests or 'experiments'
            -115 * (e.g. a random reads test, a random 
            writes test, etc.). Pass on the
            -116 * command-line which test to run and how 
            many clients are participating in
            -117 * this experiment. Run {@code 
            PerformanceEvaluation --help} to obtain usage.
            -118 *
            -119 * 

            This class sets up and runs the evaluation programs described in -120 * Section 7, Performance Evaluation, of the Bigtable; -122 * paper, pages 8-10. -123 * -124 *

            By default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.fi


            [27/51] [partial] hbase-site git commit: Published site at 87f5b5f3411d96c31b4cb61b9a57ced22be91d1f.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html
            index e1bc325..63e7421 100644
            --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html
            +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html
            @@ -66,5125 +66,5224 @@
             058import 
            java.util.concurrent.TimeoutException;
             059import 
            java.util.concurrent.atomic.AtomicBoolean;
             060import 
            java.util.concurrent.atomic.AtomicInteger;
            -061import org.apache.commons.io.IOUtils;
            -062import 
            org.apache.commons.lang3.RandomStringUtils;
            -063import 
            org.apache.commons.lang3.StringUtils;
            -064import 
            org.apache.hadoop.conf.Configuration;
            -065import 
            org.apache.hadoop.conf.Configured;
            -066import 
            org.apache.hadoop.fs.FSDataOutputStream;
            -067import org.apache.hadoop.fs.FileStatus;
            -068import org.apache.hadoop.fs.FileSystem;
            -069import org.apache.hadoop.fs.Path;
            -070import 
            org.apache.hadoop.fs.permission.FsAction;
            -071import 
            org.apache.hadoop.fs.permission.FsPermission;
            -072import 
            org.apache.hadoop.hbase.Abortable;
            -073import org.apache.hadoop.hbase.Cell;
            -074import 
            org.apache.hadoop.hbase.CellUtil;
            -075import 
            org.apache.hadoop.hbase.ClusterMetrics;
            -076import 
            org.apache.hadoop.hbase.ClusterMetrics.Option;
            -077import 
            org.apache.hadoop.hbase.HBaseConfiguration;
            -078import 
            org.apache.hadoop.hbase.HBaseInterfaceAudience;
            -079import 
            org.apache.hadoop.hbase.HConstants;
            -080import 
            org.apache.hadoop.hbase.HRegionInfo;
            -081import 
            org.apache.hadoop.hbase.HRegionLocation;
            -082import 
            org.apache.hadoop.hbase.KeyValue;
            -083import 
            org.apache.hadoop.hbase.MasterNotRunningException;
            -084import 
            org.apache.hadoop.hbase.MetaTableAccessor;
            -085import 
            org.apache.hadoop.hbase.RegionLocations;
            -086import 
            org.apache.hadoop.hbase.ServerName;
            -087import 
            org.apache.hadoop.hbase.TableName;
            -088import 
            org.apache.hadoop.hbase.ZooKeeperConnectionException;
            -089import 
            org.apache.hadoop.hbase.client.Admin;
            -090import 
            org.apache.hadoop.hbase.client.ClusterConnection;
            -091import 
            org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
            -092import 
            org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
            -093import 
            org.apache.hadoop.hbase.client.Connection;
            -094import 
            org.apache.hadoop.hbase.client.ConnectionFactory;
            -095import 
            org.apache.hadoop.hbase.client.Delete;
            -096import 
            org.apache.hadoop.hbase.client.Get;
            -097import 
            org.apache.hadoop.hbase.client.Put;
            -098import 
            org.apache.hadoop.hbase.client.RegionInfo;
            -099import 
            org.apache.hadoop.hbase.client.RegionInfoBuilder;
            -100import 
            org.apache.hadoop.hbase.client.RegionReplicaUtil;
            -101import 
            org.apache.hadoop.hbase.client.Result;
            -102import 
            org.apache.hadoop.hbase.client.RowMutations;
            -103import 
            org.apache.hadoop.hbase.client.Table;
            -104import 
            org.apache.hadoop.hbase.client.TableDescriptor;
            -105import 
            org.apache.hadoop.hbase.client.TableDescriptorBuilder;
            -106import 
            org.apache.hadoop.hbase.client.TableState;
            -107import 
            org.apache.hadoop.hbase.io.FileLink;
            -108import 
            org.apache.hadoop.hbase.io.HFileLink;
            -109import 
            org.apache.hadoop.hbase.io.hfile.CacheConfig;
            -110import 
            org.apache.hadoop.hbase.io.hfile.HFile;
            -111import 
            org.apache.hadoop.hbase.log.HBaseMarkers;
            -112import 
            org.apache.hadoop.hbase.master.MasterFileSystem;
            -113import 
            org.apache.hadoop.hbase.master.RegionState;
            -114import 
            org.apache.hadoop.hbase.regionserver.HRegion;
            -115import 
            org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
            -116import 
            org.apache.hadoop.hbase.regionserver.StoreFileInfo;
            -117import 
            org.apache.hadoop.hbase.replication.ReplicationException;
            -118import 
            org.apache.hadoop.hbase.security.AccessDeniedException;
            -119import 
            org.apache.hadoop.hbase.security.UserProvider;
            -120import 
            org.apache.hadoop.hbase.util.Bytes.ByteArrayComparator;
            -121import 
            org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE;
            -122import 
            org.apache.hadoop.hbase.util.hbck.HFileCorruptionChecker;
            -123import 
            org.apache.hadoop.hbase.util.hbck.ReplicationChecker;
            -124import 
            org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandler;
            -125import 
            org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandlerImpl;
            -126import org.apache.hadoop.hbase.wal.WAL;
            -127import 
            org.apache.hadoop.hbase.wal.WALFactory;
            -128import 
            org.apache.hadoop.hbase.wal.WALSplitter;
            -129import 
            org.apache.hadoop.hbase.zookeeper.MetaTableLocator;
            -130import 
            org.apache.hadoop.hbase.zookeeper.ZKUtil;
            -131import 
            org.apache.hadoop.hbase.zookeeper.ZKWatcher;
            -132import 
            org.apache.hadoop.hbase.zookeeper.ZNodePaths;
            -133import 
            org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException;
            -134import 
            org.apache.hadoop.ipc.RemoteException;
            -135import 
            org.apache.hadoop.security.UserGroupInformation;
            -136import 
            org.apach

            [27/51] [partial] hbase-site git commit: Published site at 2912c953551bedbfbf30c32c156ed7bb187d54c3.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html
            index 56ec3ac..8f4cacb 100644
            --- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html
            +++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html
            @@ -229,9 +229,9 @@
             221   * Note that we don't document 
            --expected-count, because it's intended for test.
             222   */
             223  private static void printUsage() {
            -224System.err.println("Usage: RowCounter 
            [options]  " +
            -225"[--starttime=[start] 
            --endtime=[end] " +
            -226
            "[--range=[startKey],[endKey][;[startKey],[endKey]...]] [ 
            ...]");
            +224System.err.println("Usage: hbase 
            rowcounter [options]  "
            +225+ "[--starttime= 
            --endtime=] "
            +226+ 
            "[--range=[startKey],[endKey][;[startKey],[endKey]...]] [ 
            ...]");
             227System.err.println("For performance 
            consider the following options:\n"
             228+ 
            "-Dhbase.client.scanner.caching=100\n"
             229+ 
            "-Dmapreduce.map.speculative=false");
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
            index a5477ac..9501e97 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
            @@ -357,46 +357,51 @@
             349  System.err.println("ERROR: " + 
            errorMsg);
             350}
             351System.err.println("Usage: " + NAME + 
            " [options]   []");
            -352System.err.println("Read all WAL 
            entries for .");
            -353System.err.println("If no tables 
            (\"\") are specific, all tables are imported.");
            -354System.err.println("(Careful, even 
            hbase:meta entries will be imported" + " in that case.)");
            -355System.err.println("Otherwise 
             is a comma separated list of tables.\n");
            -356System.err.println("The WAL entries 
            can be mapped to new set of tables via .");
            -357
            System.err.println(" is a command separated list of 
            targettables.");
            +352System.err.println("Replay all WAL 
            files into HBase.");
            +353System.err.println(" is 
            a comma separated list of tables.");
            +354System.err.println("If no tables 
            (\"\") are specified, all tables are imported.");
            +355System.err.println("(Be careful, 
            hbase:meta entries will be imported in this case.)\n");
            +356System.err.println("WAL entries can 
            be mapped to new set of tables via .");
            +357
            System.err.println(" is a comma separated list of target 
            tables.");
             358System.err.println("If specified, 
            each table in  must have a mapping.\n");
             359System.err.println("By default " + 
            NAME + " will load data directly into HBase.");
            -360System.err.println("To generate 
            HFiles for a bulk data load instead, pass the option:");
            +360System.err.println("To generate 
            HFiles for a bulk data load instead, pass the following option:");
             361System.err.println("  -D" + 
            BULK_OUTPUT_CONF_KEY + "=/path/for/output");
             362System.err.println("  (Only one table 
            can be specified, and no mapping is allowed!)");
            -363System.err.println("Other options: 
            (specify time range to WAL edit to consider)");
            +363System.err.println("Time range 
            options:");
             364System.err.println("  -D" + 
            WALInputFormat.START_TIME_KEY + "=[date|ms]");
             365System.err.println("  -D" + 
            WALInputFormat.END_TIME_KEY + "=[date|ms]");
            -366System.err.println("   -D " + 
            JOB_NAME_CONF_KEY
            -367+ "=jobName - use the specified 
            mapreduce job name for the wal player");
            -368System.err.println("For performance 
            also consider the following options:\n"
            -369+ "  
            -Dmapreduce.map.speculative=false\n" + "  
            -Dmapreduce.reduce.speculative=false");
            -370  }
            -371
            -372  /**
            -373   * Main entry point.
            -374   * @param args The command line 
            parameters.
            -375   * @throws Exception When running the 
            job fails.
            -376   */
            -377  public static void main(String[] args) 
            throws Exception {
            -378int ret = ToolRunner.run(new 
            WALPlayer(HBaseConfiguration.create()), args);
            -379System.exit(ret);
            -380  }
            -381
            -382  @Override
            -383  public int run(String[] args) throws 
            Exception {
            -384if (args.length < 2) {
            -385  usage("Wrong number of arguments: " 
            + args.length);
            -386  System.exit(-1);

            [27/51] [partial] hbase-site git commit: Published site at 2a2258656b2fcd92b967131b6c1f037363553bc4.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html 
            b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html
            index efc1791..6e81af2 100644
            --- a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html
            +++ b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html
            @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private static class RegionMover.MoveWithoutAck
            +private static class RegionMover.MoveWithoutAck
             extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object
             implements https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true";
             title="class or interface in java.util.concurrent">CallableBoolean>
             Move Regions without Acknowledging.Usefule in case of RS 
            shutdown as we might want to shut the
            @@ -228,7 +228,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
             
             
             admin
            -private Admin admin
            +private Admin admin
             
             
             
            @@ -237,7 +237,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
             
             
             region
            -private RegionInfo region
            +private RegionInfo region
             
             
             
            @@ -246,7 +246,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
             
             
             targetServer
            -private https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String targetServer
            +private https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String targetServer
             
             
             
            @@ -255,7 +255,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
             
             
             movedRegions
            -private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List movedRegions
            +private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List movedRegions
             
             
             
            @@ -264,7 +264,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
             
             
             sourceServer
            -private https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String sourceServer
            +private https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String sourceServer
             
             
             
            @@ -281,7 +281,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
             
             
             MoveWithoutAck
            -public MoveWithoutAck(Admin admin,
            +public MoveWithoutAck(Admin admin,
               RegionInfo regionInfo,
               https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String sourceServer,
               https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String targetServer,
            @@ -302,7 +302,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
             
             
             call
            -public https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true";
             title="class or interface in java.lang">Boolean call()
            +public https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true";
             title="class or interface in java.lang">Boolean call()
             
             Specified by:
             https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true#call--";
             title="class or interface in java.util.concurrent">call in 
            interface https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true";
             title="class or interface in java.util.concurrent">CallableBoolean>
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html 
            b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html
            index c451701..d4d8db5 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html
            @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -public static class RegionMover.RegionMoverBuilder
            +pub

            [27/51] [partial] hbase-site git commit: Published site at e468b4022f76688851b3e0c34722f01a56bd624f.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
            --
            diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html 
            b/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
            index dbb2f3e..c887b35 100644
            --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
            +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
            @@ -36,843 +36,849 @@
             028import 
            org.apache.hadoop.hbase.CompareOperator;
             029import 
            org.apache.hadoop.hbase.HTableDescriptor;
             030import 
            org.apache.hadoop.hbase.TableName;
            -031import 
            org.apache.yetus.audience.InterfaceAudience;
            -032import 
            org.apache.hadoop.hbase.client.coprocessor.Batch;
            -033import 
            org.apache.hadoop.hbase.filter.CompareFilter;
            -034import 
            org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
            -035
            -036import com.google.protobuf.Descriptors;
            -037import com.google.protobuf.Message;
            -038import com.google.protobuf.Service;
            -039import 
            com.google.protobuf.ServiceException;
            -040
            -041/**
            -042 * Used to communicate with a single 
            HBase table.
            -043 * Obtain an instance from a {@link 
            Connection} and call {@link #close()} afterwards.
            -044 *
            -045 * 

            Table can be used to get, put, delete or scan data from a table. -046 * @see ConnectionFactory -047 * @see Connection -048 * @see Admin -049 * @see RegionLocator -050 * @since 0.99.0 -051 */ -052@InterfaceAudience.Public -053public interface Table extends Closeable { -054 /** -055 * Gets the fully qualified table name instance of this table. -056 */ -057 TableName getName(); -058 -059 /** -060 * Returns the {@link org.apache.hadoop.conf.Configuration} object used by this instance. -061 *

            -062 * The reference returned is not a copy, so any change made to it will -063 * affect this instance. -064 */ -065 Configuration getConfiguration(); -066 -067 /** -068 * Gets the {@link org.apache.hadoop.hbase.HTableDescriptor table descriptor} for this table. -069 * @throws java.io.IOException if a remote or network exception occurs. -070 * @deprecated since 2.0 version and will be removed in 3.0 version. -071 * use {@link #getDescriptor()} -072 */ -073 @Deprecated -074 HTableDescriptor getTableDescriptor() throws IOException; -075 -076 /** -077 * Gets the {@link org.apache.hadoop.hbase.client.TableDescriptor table descriptor} for this table. -078 * @throws java.io.IOException if a remote or network exception occurs. -079 */ -080 TableDescriptor getDescriptor() throws IOException; -081 -082 /** -083 * Test for the existence of columns in the table, as specified by the Get. -084 *

            -085 * -086 * This will return true if the Get matches one or more keys, false if not. -087 *

            -088 * -089 * This is a server-side call so it prevents any data from being transfered to -090 * the client. -091 * -092 * @param get the Get -093 * @return true if the specified Get matches one or more keys, false if not -094 * @throws IOException e -095 */ -096 boolean exists(Get get) throws IOException; -097 -098 /** -099 * Test for the existence of columns in the table, as specified by the Gets. -100 *

            -101 * -102 * This will return an array of booleans. Each value will be true if the related Get matches -103 * one or more keys, false if not. -104 *

            -105 * -106 * This is a server-side call so it prevents any data from being transferred to -107 * the client. -108 * -109 * @param gets the Gets -110 * @return Array of boolean. True if the specified Get matches one or more keys, false if not. -111 * @throws IOException e -112 */ -113 boolean[] exists(List gets) throws IOException; -114 -115 /** -116 * Test for the existence of columns in the table, as specified by the Gets. -117 * This will return an array of booleans. Each value will be true if the related Get matches -118 * one or more keys, false if not. -119 * This is a server-side call so it prevents any data from being transferred to -120 * the client. -121 * -122 * @param gets the Gets -123 * @return Array of boolean. True if the specified Get matches one or more keys, false if not. -124 * @throws IOException e -125 * @deprecated since 2.0 version and will be removed in 3.0 version. -126 * use {@link #exists(List)} -127 */ -128 @Deprecated -129 default boolean[] existsAll(List gets) throws IOException { -130return exists(gets); -131 } -132 -133 /** -134 * Method that does a batch call on Deletes, Gets, Puts, Increments, Appends, RowMutations. -135 * The ordering of execution of the actions is not defined. Meaning if you do a Put and a -136 * Get in the same {@link #batch} call, you will not necessarily be -137 * guaranteed that the Get returns what the Put had put. -138 * -139 * @param actions list of Get, Put, Delete, Increment, A


            [27/51] [partial] hbase-site git commit: Published site at 64061f896fe21512504e3886a400759e88b519da.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            index 5360e0a..d4296e6 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            @@ -627,13 +627,18 @@ service.
            https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true";
             title="class or interface in 
            java.util.concurrent">ExecutorService pool) 
             
             
            +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">ListString,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true";
             title="class or interface in java.lang">Long>>
            +MetaTableAccessor.getTableEncodedRegionNameAndLastBarrier(Connection conn,
            +   TableName tableName) 
            +
            +
             static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             MetaTableAccessor.getTableRegions(Connection connection,
            TableName tableName)
             Gets all of the regions of the specified table.
             
             
            -
            +
             static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             MetaTableAccessor.getTableRegions(Connection connection,
            TableName tableName,
            @@ -641,14 +646,14 @@ service.
             Gets all of the regions of the specified table.
             
             
            -
            +
             static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List>
             MetaTableAccessor.getTableRegionsAndLocations(Connection connection,
            TableName tableName)
             Do not use this method to get meta table regions, use 
            methods in MetaTableLocator instead.
             
             
            -
            +
             static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List>
             MetaTableAccessor.getTableRegionsAndLocations(Connection connection,
            TableName tableName,
            @@ -656,37 +661,37 @@ service.
             Do not use this method to get meta table regions, use 
            methods in MetaTableLocator instead.
             
             
            -
            +
             static byte[]
             MetaTableAccessor.getTableStartRowForMeta(TableName tableName,
            MetaTableAccessor.QueryType type) 
             
            -
            +
             static https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureOptional>
             AsyncMetaTableAccessor.getTableState(AsyncTable metaTable,
              TableName tableName) 
             
            -
            +
             static TableState
             MetaTableAccessor.getTableState(Connection conn,
              TableName tableName)
             Fetch table state for given table from META table
             
             
            -
            +
             static byte[]
             MetaTableAccessor.getTableStopRowForMeta(TableName tableName,
               MetaTableAccessor.QueryType type) 
             
            -
            +
             static boolean
             TableName.isMetaTableName(TableName tn) 
             
            -
            +
             TableDescriptor
             TableDescriptors.remove(TableName tablename) 
             
            -
            +
             static void
             MetaTableAccessor.scanMeta(Connection connection,
             MetaTableAccessor.Visitor visitor,
            @@ -697,7 +702,7 @@ service.
              given row.
             
             
            -
            +
             static void
             MetaTableAccessor.scanMeta(Connection connection,
             TableName table,
            @@ -705,25 +710,25 @@ service.
             int maxRows,
             MetaTableAccessor.Visitor visitor) 
             
            -
            +
             static void
             MetaTableAccessor.scanMetaForTableRegions(Connection connection,
            MetaTableAccessor.Visitor visitor,
            TableName tableName) 
             
            -
            +
             static https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureBoolean>
             AsyncMetaTableAccessor.tableExists(AsyncTable metaTable,
            TableName tableName) 
             
            -
            +
             static boolean
             MetaTableAccessor.tableExists(Connection connection,
            TableName tableName)
             Checks if the specified table exists.
             
             
            -
            +
             static void
             MetaTableAccessor.updateTableState(Connection conn,
             TableName tableName,
            @@ -2052,119 +2057,119 @@ service.
             
             
             private TableName
            -SnapshotDescri

            [27/51] [partial] hbase-site git commit: Published site at 4cb40e6d846ce1f28ffb40d388c9efb753197813.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/class-use/Size.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Size.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/Size.html
            index 3f1b032..e8d1010 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/Size.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Size.html
            @@ -189,130 +189,130 @@
             
             
             Size
            -RegionLoad.getBloomFilterSize()
            -Deprecated. 
            - 
            -
            -
            -Size
             RegionMetrics.getBloomFilterSize() 
             
            -
            -Size
            -RegionMetricsBuilder.RegionMetricsImpl.getBloomFilterSize() 
            -
             
             Size
            -ServerLoad.getMaxHeapSize()
            +RegionLoad.getBloomFilterSize()
             Deprecated. 
              
             
             
             Size
            -ServerMetrics.getMaxHeapSize() 
            +RegionMetricsBuilder.RegionMetricsImpl.getBloomFilterSize() 
             
             
             Size
            -ServerMetricsBuilder.ServerMetricsImpl.getMaxHeapSize() 
            +ServerMetrics.getMaxHeapSize() 
             
             
             Size
            -RegionLoad.getMemStoreSize()
            +ServerLoad.getMaxHeapSize()
             Deprecated. 
              
             
             
             Size
            -RegionMetrics.getMemStoreSize() 
            +ServerMetricsBuilder.ServerMetricsImpl.getMaxHeapSize() 
             
             
             Size
            -RegionMetricsBuilder.RegionMetricsImpl.getMemStoreSize() 
            +RegionMetrics.getMemStoreSize() 
             
             
             Size
            -RegionLoad.getStoreFileIndexSize()
            +RegionLoad.getMemStoreSize()
             Deprecated. 
              
             
             
             Size
            +RegionMetricsBuilder.RegionMetricsImpl.getMemStoreSize() 
            +
            +
            +Size
             RegionMetrics.getStoreFileIndexSize()
             TODO: why we pass the same value to different counters? 
            Currently, the value from
              getStoreFileIndexSize() is same with getStoreFileRootLevelIndexSize()
              see HRegionServer#createRegionLoad.
             
             
            -
            -Size
            -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileIndexSize() 
            -
             
             Size
            -RegionLoad.getStoreFileRootLevelIndexSize()
            +RegionLoad.getStoreFileIndexSize()
             Deprecated. 
              
             
             
             Size
            -RegionMetrics.getStoreFileRootLevelIndexSize() 
            +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileIndexSize() 
             
             
             Size
            -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileRootLevelIndexSize() 
            +RegionMetrics.getStoreFileRootLevelIndexSize() 
             
             
             Size
            -RegionLoad.getStoreFileSize()
            +RegionLoad.getStoreFileRootLevelIndexSize()
             Deprecated. 
              
             
             
             Size
            -RegionMetrics.getStoreFileSize() 
            +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileRootLevelIndexSize() 
             
             
             Size
            -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileSize() 
            +RegionMetrics.getStoreFileSize() 
             
             
             Size
            -RegionLoad.getStoreFileUncompressedDataIndexSize()
            +RegionLoad.getStoreFileSize()
             Deprecated. 
              
             
             
             Size
            -RegionMetrics.getStoreFileUncompressedDataIndexSize() 
            +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileSize() 
             
             
             Size
            -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileUncompressedDataIndexSize() 
            +RegionMetrics.getStoreFileUncompressedDataIndexSize() 
             
             
             Size
            -RegionLoad.getUncompressedStoreFileSize()
            +RegionLoad.getStoreFileUncompressedDataIndexSize()
             Deprecated. 
              
             
             
             Size
            -RegionMetrics.getUncompressedStoreFileSize() 
            +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileUncompressedDataIndexSize() 
             
             
             Size
            -RegionMetricsBuilder.RegionMetricsImpl.getUncompressedStoreFileSize() 
            +RegionMetrics.getUncompressedStoreFileSize() 
             
             
             Size
            -ServerLoad.getUsedHeapSize()
            +RegionLoad.getUncompressedStoreFileSize()
             Deprecated. 
              
             
             
             Size
            +RegionMetricsBuilder.RegionMetricsImpl.getUncompressedStoreFileSize() 
            +
            +
            +Size
             ServerMetrics.getUsedHeapSize() 
             
            +
            +Size
            +ServerLoad.getUsedHeapSize()
            +Deprecated. 
            + 
            +
             
             Size
             ServerMetricsBuilder.ServerMetricsImpl.getUsedHeapSize() 
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html
            index 63833f7..72d579d 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html
            @@ -122,11 +122,11 @@
             
             
             TableDescriptors
            -HMaster.getTableDescriptors() 
            +MasterServices.getTableDescriptors() 
             
             
             TableDescriptors
            -MasterServices.getTableDescriptors() 
            +HMaster.getTableDescriptors() 
             
             
             
            
            
            

            [27/51] [partial] hbase-site git commit: Published site at 8ab7b20f48951d77945181024f5e15842bc253c4.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
            index ecf500c..0cd5a4e 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
            @@ -238,8355 +238,8368 @@
             230  public static final String 
            HBASE_MAX_CELL_SIZE_KEY = "hbase.server.keyvalue.maxsize";
             231  public static final int 
            DEFAULT_MAX_CELL_SIZE = 10485760;
             232
            -233  public static final String 
            HBASE_REGIONSERVER_MINIBATCH_SIZE =
            -234  
            "hbase.regionserver.minibatch.size";
            -235  public static final int 
            DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2;
            -236
            -237  /**
            -238   * This is the global default value for 
            durability. All tables/mutations not
            -239   * defining a durability or using 
            USE_DEFAULT will default to this value.
            -240   */
            -241  private static final Durability 
            DEFAULT_DURABILITY = Durability.SYNC_WAL;
            +233  /**
            +234   * This is the global default value for 
            durability. All tables/mutations not
            +235   * defining a durability or using 
            USE_DEFAULT will default to this value.
            +236   */
            +237  private static final Durability 
            DEFAULT_DURABILITY = Durability.SYNC_WAL;
            +238
            +239  public static final String 
            HBASE_REGIONSERVER_MINIBATCH_SIZE =
            +240  
            "hbase.regionserver.minibatch.size";
            +241  public static final int 
            DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2;
             242
            -243  final AtomicBoolean closed = new 
            AtomicBoolean(false);
            -244
            -245  /* Closing can take some time; use the 
            closing flag if there is stuff we don't
            -246   * want to do while in closing state; 
            e.g. like offer this region up to the
            -247   * master as a region to close if the 
            carrying regionserver is overloaded.
            -248   * Once set, it is never cleared.
            -249   */
            -250  final AtomicBoolean closing = new 
            AtomicBoolean(false);
            -251
            -252  /**
            -253   * The max sequence id of flushed data 
            on this region. There is no edit in memory that is
            -254   * less that this sequence id.
            -255   */
            -256  private volatile long maxFlushedSeqId = 
            HConstants.NO_SEQNUM;
            -257
            -258  /**
            -259   * Record the sequence id of last flush 
            operation. Can be in advance of
            -260   * {@link #maxFlushedSeqId} when 
            flushing a single column family. In this case,
            -261   * {@link #maxFlushedSeqId} will be 
            older than the oldest edit in memory.
            -262   */
            -263  private volatile long lastFlushOpSeqId 
            = HConstants.NO_SEQNUM;
            -264
            -265  /**
            -266   * The sequence id of the last replayed 
            open region event from the primary region. This is used
            -267   * to skip entries before this due to 
            the possibility of replay edits coming out of order from
            -268   * replication.
            -269   */
            -270  protected volatile long 
            lastReplayedOpenRegionSeqId = -1L;
            -271  protected volatile long 
            lastReplayedCompactionSeqId = -1L;
            -272
            -273  
            //
            -274  // Members
            -275  
            //
            -276
            -277  // map from a locked row to the context 
            for that lock including:
            -278  // - CountDownLatch for threads waiting 
            on that row
            -279  // - the thread that owns the lock 
            (allow reentrancy)
            -280  // - reference count of (reentrant) 
            locks held by the thread
            -281  // - the row itself
            -282  private final 
            ConcurrentHashMap lockedRows =
            -283  new ConcurrentHashMap<>();
            -284
            -285  protected final Map stores =
            -286  new 
            ConcurrentSkipListMap<>(Bytes.BYTES_RAWCOMPARATOR);
            +243  public static final String 
            WAL_HSYNC_CONF_KEY = "hbase.wal.hsync";
            +244  public static final boolean 
            DEFAULT_WAL_HSYNC = false;
            +245
            +246  final AtomicBoolean closed = new 
            AtomicBoolean(false);
            +247
            +248  /* Closing can take some time; use the 
            closing flag if there is stuff we don't
            +249   * want to do while in closing state; 
            e.g. like offer this region up to the
            +250   * master as a region to close if the 
            carrying regionserver is overloaded.
            +251   * Once set, it is never cleared.
            +252   */
            +253  final AtomicBoolean closing = new 
            AtomicBoolean(false);
            +254
            +255  /**
            +256   * The max sequence id of flushed data 
            on this region. There is no edit in memory that is
            +257   * less that this sequence id.
            +258   */
            +259  private volatile long maxFlushedSeqId = 
            HConstants.NO_SEQNUM;
            +260
            +261  /**
            +262   * Record the sequence id of last flush 
            operation. Can be in advance of
            +263   * {@link #maxFlushedSeqId} when 
            flushing a single column family. In this case,
            +264   * {@link #maxFlushedSeqId} will be 
            older than the oldest edit in memory.
            +265   */
            +266  private vol

            [27/51] [partial] hbase-site git commit: Published site at 00095a2ef9442e3fd86c04876c9d91f2f8b23ad8.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
            index f47d627..c3d225c 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
            @@ -117,219 +117,219 @@
             109   */
             110  public static boolean 
            archiveRegion(FileSystem fs, Path rootdir, Path tableDir, Path regionDir)
             111  throws IOException {
            -112if (LOG.isDebugEnabled()) {
            -113  LOG.debug("ARCHIVING " + 
            regionDir.toString());
            -114}
            -115
            -116// otherwise, we archive the files
            -117// make sure we can archive
            -118if (tableDir == null || regionDir == 
            null) {
            -119  LOG.error("No archive directory 
            could be found because tabledir (" + tableDir
            -120  + ") or regiondir (" + 
            regionDir + "was null. Deleting files instead.");
            -121  deleteRegionWithoutArchiving(fs, 
            regionDir);
            -122  // we should have archived, but 
            failed to. Doesn't matter if we deleted
            -123  // the archived files correctly or 
            not.
            -124  return false;
            -125}
            -126
            -127// make sure the regiondir lives 
            under the tabledir
            -128
            Preconditions.checkArgument(regionDir.toString().startsWith(tableDir.toString()));
            -129Path regionArchiveDir = 
            HFileArchiveUtil.getRegionArchiveDir(rootdir,
            -130FSUtils.getTableName(tableDir),
            -131regionDir.getName());
            -132
            -133FileStatusConverter getAsFile = new 
            FileStatusConverter(fs);
            -134// otherwise, we attempt to archive 
            the store files
            -135
            -136// build collection of just the store 
            directories to archive
            -137Collection toArchive = 
            new ArrayList<>();
            -138final PathFilter dirFilter = new 
            FSUtils.DirFilter(fs);
            -139PathFilter nonHidden = new 
            PathFilter() {
            -140  @Override
            -141  public boolean accept(Path file) 
            {
            -142return dirFilter.accept(file) 
            && !file.getName().toString().startsWith(".");
            -143  }
            -144};
            -145FileStatus[] storeDirs = 
            FSUtils.listStatus(fs, regionDir, nonHidden);
            -146// if there no files, we can just 
            delete the directory and return;
            -147if (storeDirs == null) {
            -148  LOG.debug("Region directory " + 
            regionDir + " empty.");
            -149  return 
            deleteRegionWithoutArchiving(fs, regionDir);
            -150}
            -151
            -152// convert the files in the region to 
            a File
            -153
            toArchive.addAll(Lists.transform(Arrays.asList(storeDirs), getAsFile));
            -154LOG.debug("Archiving " + 
            toArchive);
            -155List failedArchive = 
            resolveAndArchive(fs, regionArchiveDir, toArchive,
            -156
            EnvironmentEdgeManager.currentTime());
            -157if (!failedArchive.isEmpty()) {
            -158  throw new 
            FailedArchiveException("Failed to archive/delete all the files for region:"
            -159  + regionDir.getName() + " into 
            " + regionArchiveDir
            -160  + ". Something is probably awry 
            on the filesystem.",
            -161  
            Collections2.transform(failedArchive, FUNC_FILE_TO_PATH));
            -162}
            -163// if that was successful, then we 
            delete the region
            -164return 
            deleteRegionWithoutArchiving(fs, regionDir);
            -165  }
            -166
            -167  /**
            -168   * Remove from the specified region the 
            store files of the specified column family,
            -169   * either by archiving them or outright 
            deletion
            -170   * @param fs the filesystem where the 
            store files live
            -171   * @param conf {@link Configuration} to 
            examine to determine the archive directory
            -172   * @param parent Parent region hosting 
            the store files
            -173   * @param tableDir {@link Path} to 
            where the table is being stored (for building the archive path)
            -174   * @param family the family hosting the 
            store files
            -175   * @throws IOException if the files 
            could not be correctly disposed.
            -176   */
            -177  public static void 
            archiveFamily(FileSystem fs, Configuration conf,
            -178  RegionInfo parent, Path tableDir, 
            byte[] family) throws IOException {
            -179Path familyDir = new Path(tableDir, 
            new Path(parent.getEncodedName(), Bytes.toString(family)));
            -180archiveFamilyByFamilyDir(fs, conf, 
            parent, familyDir, family);
            -181  }
            -182
            -183  /**
            -184   * Removes from the specified region 
            the store files of the specified column family,
            -185   * either by archiving them or outright 
            deletion
            -186   * @param fs the filesystem where the 
            store files live
            -187   * @param conf {@link Configuration} to 
            examine to determine the archive directory
            -188   * @param parent Parent region hosting 
            the store files
            -189   * @param familyDir {@link Path} to 
            where the family is being stored
            -190   * @param family the family hosting the 
            store files
            -191   * @throws IOException if the files 
            could not be correctly disposed.
            -192 

            [27/51] [partial] hbase-site git commit: Published site at 22f4def942f8a3367d0ca6598317e9b9a7d0cfcd.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
             
            b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
            index 067b5c4..beb9075 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
            @@ -126,7 +126,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private static class RegionCoprocessorHost.RegionEnvironment
            +private static class RegionCoprocessorHost.RegionEnvironment
             extends BaseEnvironment
             implements RegionCoprocessorEnvironment
             Encapsulation of the environment of each coprocessor
            @@ -303,7 +303,7 @@ implements 
             
             region
            -private Region region
            +private Region region
             
             
             
            @@ -312,7 +312,7 @@ implements 
             
             sharedData
            -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true";
             title="class or interface in java.util.concurrent">ConcurrentMapString,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object> sharedData
            +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true";
             title="class or interface in java.util.concurrent">ConcurrentMapString,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object> sharedData
             
             
             
            @@ -321,7 +321,7 @@ implements 
             
             metricRegistry
            -private final MetricRegistry metricRegistry
            +private final MetricRegistry metricRegistry
             
             
             
            @@ -330,7 +330,7 @@ implements 
             
             services
            -private final RegionServerServices services
            +private final RegionServerServices services
             
             
             
            @@ -347,7 +347,7 @@ implements 
             
             RegionEnvironment
            -public RegionEnvironment(RegionCoprocessor impl,
            +public RegionEnvironment(RegionCoprocessor impl,
              int priority,
              int seq,
              org.apache.hadoop.conf.Configuration conf,
            @@ -376,7 +376,7 @@ implements 
             
             getRegion
            -public Region getRegion()
            +public Region getRegion()
             
             Specified by:
             getRegion in
             interface RegionCoprocessorEnvironment
            @@ -391,7 +391,7 @@ implements 
             
             getOnlineRegions
            -public OnlineRegions getOnlineRegions()
            +public OnlineRegions getOnlineRegions()
             
             Specified by:
             getOnlineRegions in
             interface RegionCoprocessorEnvironment
            @@ -406,7 +406,7 @@ implements 
             
             getConnection
            -public Connection getConnection()
            +public Connection getConnection()
             Description copied from 
            interface: RegionCoprocessorEnvironment
             Returns the hosts' Connection to the Cluster. Do not 
            close! This is a shared connection
              with the hosting server. Throws https://docs.oracle.com/javase/8/docs/api/java/lang/UnsupportedOperationException.html?is-external=true";
             title="class or interface in 
            java.lang">UnsupportedOperationException if you try to close
            @@ -445,7 +445,7 @@ implements 
             
             createConnection
            -public Connection createConnection(org.apache.hadoop.conf.Configuration conf)
            +public Connection createConnection(org.apache.hadoop.conf.Configuration conf)
             throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in java.io">IOException
             Description copied from 
            interface: RegionCoprocessorEnvironment
             Creates a cluster connection using the passed Configuration.
            @@ -481,7 +481,7 @@ implements 
             
             getServerName
            -public ServerName getServerName()
            +public ServerName getServerName()
             
             Specified by:
             getServerName in
             interface RegionCoprocessorEnvironment
            @@ -496,7 +496,7 @@ implements 
             
             shutdown
            -public void shutdown()
            +public void shutdown()
             Description copied from 
            class: BaseEnvironment
             Clean up the environment
             
            @@ -511,7 +511,7 @@ implements 
             
             getSharedData
            -public https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true";
             title="class or interface in java.util.concurrent">ConcurrentMapString,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object> getSharedData()
            +public https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true";
             title="class or inte

            [27/51] [partial] hbase-site git commit: Published site at 31da4d0bce69b3a47066a5df675756087ce4dc60.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
             
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
            index fa987ee..a13341f 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
            @@ -317,7 +317,7 @@ extends AbstractStateMachineTableProcedure
            -acquireLock,
             checkTableModifiable,
             getRegionDir,
             getUser,
             releaseLock,
             <
             a 
            href="../../../../../../org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html#releaseSyncLatch--">releaseSyncLatch,
             setUser
            +acquireLock,
             checkTableModifiable,
             getRegionDir,
             getUser,
             preflightChecks, releaseLock,
             releaseSyncLatch,
             setUser
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html 
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html
            index 72ff612..671b283 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html
            @@ -305,7 +305,7 @@ extends AbstractStateMachineTableProcedure
            -getRegionDir,
             getUser,
             releaseSyncLatch,
             setUser
            +getRegionDir,
             getUser,
             preflightChecks,
             releaseSyncLatch,
             setUser
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
             
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
            index 347cea7..2881954 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
            @@ -129,11 +129,13 @@ var activeTableTab = "activeTableTab";
             
             
             @InterfaceAudience.Private
            -public class MergeTableRegionsProcedure
            +public class MergeTableRegionsProcedure
             extends AbstractStateMachineTableProcedure
             The procedure to Merge a region in a table.
              This procedure takes an exclusive table lock since it is working over 
            multiple regions.
            - It holds the lock for the life of the procedure.
            + It holds the lock for the life of the procedure.
            + Throws exception on construction if determines context hostile to merge 
            (cluster going
            + down or master is shutting down or table is disabled).
             
             
             
            @@ -520,7 +522,7 @@ extends AbstractStateMachineTableProcedure
            -checkTableModifiable,
             getRegionDir,
             getUser,
             releaseSyncLatch,
             setUser
            +checkTableModifiable,
             getRegionDir,
             getUser,
             preflightChecks,
             releaseSyncLatch,
             setUser
             
             
             
            @@ -563,7 +565,7 @@ extends 
             
             LOG
            -private static final org.slf4j.Logger LOG
            +private static final org.slf4j.Logger LOG
             
             
             
            @@ -572,7 +574,7 @@ extends 
             
             traceEnabled
            -private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true";
             title="class or interface in java.lang">Boolean traceEnabled
            +private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true";
             title="class or interface in java.lang">Boolean traceEnabled
             
             
             
            @@ -581,7 +583,7 @@ extends 
             
             lock
            -private volatile boolean lock
            +private volatile boolean lock
             
             
             
            @@ -590,7 +592,7 @@ extends 
             
             regionLocation
            -private ServerName regionLocation
            +private ServerName regionLocation
             
             
             
            @@ -599,7 +601,7 @@ extends 
             
             regionsToMerge
            -private RegionInfo[] regionsToMerge
            +private RegionInfo[] regionsToMerge
             
             
             
            @@ -608,7 +610,7 @@ extends 
             
             mergedRegion
            -private RegionInfo mergedRegion
            +private RegionInfo mergedRegion
             
             
             
            @@ -617,7 +619,7 @@ extends 
             
             forcible
            -private boolean forcible
            +private boolean forcible
             
             
             
            @@ -634,7 +636,7 @@ extends 
             
             MergeTableRegionsProcedure
            -public MergeTableRegionsProcedure()
            +public MergeTableRegionsProcedure()
             
             
             
            @@ -643,7 +645,7 @@ extends 
             
             MergeTableRegionsProcedure
            -public MergeTableRegionsProcedure(MasterProcedureEnv env,
            +public MergeTableRegionsProcedure(MasterProcedureEnv env,
               RegionInfo regionToMergeA,
               RegionInfo regionToMergeB)
            throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             ti

            [27/51] [partial] hbase-site git commit: Published site at 6b77786dfc46d25ac5bb5f1c8a4a9eb47b52a604.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
            index 71b940f..7ac76dd 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
            @@ -106,11 +106,11 @@
             
             
             private AsyncConnectionImpl
            -RawAsyncTableImpl.conn 
            +AsyncClientScanner.conn 
             
             
             private AsyncConnectionImpl
            -AsyncBatchRpcRetryingCaller.conn 
            +AsyncRpcRetryingCallerFactory.conn 
             
             
             private AsyncConnectionImpl
            @@ -118,19 +118,19 @@
             
             
             private AsyncConnectionImpl
            -RegionCoprocessorRpcChannelImpl.conn 
            +RawAsyncTableImpl.conn 
             
             
            -protected AsyncConnectionImpl
            -AsyncRpcRetryingCaller.conn 
            +private AsyncConnectionImpl
            +RegionCoprocessorRpcChannelImpl.conn 
             
             
             private AsyncConnectionImpl
            -AsyncClientScanner.conn 
            +AsyncBatchRpcRetryingCaller.conn 
             
             
            -private AsyncConnectionImpl
            -AsyncRpcRetryingCallerFactory.conn 
            +protected AsyncConnectionImpl
            +AsyncRpcRetryingCaller.conn 
             
             
             private AsyncConnectionImpl
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
            index e71ca45..d6b1759 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
            @@ -105,13 +105,13 @@
             
             
             
            -private AsyncMasterRequestRpcRetryingCaller.Callable
            -AsyncMasterRequestRpcRetryingCaller.callable 
            -
            -
             private AsyncMasterRequestRpcRetryingCaller.Callable
             AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder.callable 
             
            +
            +private AsyncMasterRequestRpcRetryingCaller.Callable
            +AsyncMasterRequestRpcRetryingCaller.callable 
            +
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
            index 5f86673..f2374f5 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
            @@ -106,11 +106,11 @@
             
             
             private AsyncProcess
            -BufferedMutatorImpl.ap 
            +HTableMultiplexer.FlushWorker.ap 
             
             
             private AsyncProcess
            -HTableMultiplexer.FlushWorker.ap 
            +BufferedMutatorImpl.ap 
             
             
             private AsyncProcess
            @@ -137,11 +137,11 @@
             
             
             AsyncProcess
            -ClusterConnection.getAsyncProcess() 
            +ConnectionImplementation.getAsyncProcess() 
             
             
             AsyncProcess
            -ConnectionImplementation.getAsyncProcess() 
            +ClusterConnection.getAsyncProcess() 
             
             
             (package private) AsyncProcess
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html
            index c610e19..9a8d746 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html
            @@ -106,11 +106,11 @@
             
             
             private AsyncRegionLocator
            -AsyncConnectionImpl.locator 
            +AsyncTableRegionLocatorImpl.locator 
             
             
             private AsyncRegionLocator
            -AsyncTableRegionLocatorImpl.locator 
            +AsyncConnectionImpl.locator 
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html
            index 447fe97..e4ecf34 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html
            @@ -126,13 +126,13 @@
             
             
             
            -(package private) AsyncRegistry
            -AsyncConnectionImpl.registry 
            -
            -
             private AsyncRegistry
             AsyncMetaRegionLocator.registry 
             
            +
            +(package private) AsyncRegistry
            +AsyncCon

            [27/51] [partial] hbase-site git commit: Published site at 1384da71375427b522b09f06862bb5d629cef52f.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            index d788050..bba78a1 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            @@ -2052,119 +2052,119 @@ service.
             
             
             private TableName
            -RegionCoprocessorRpcChannel.table 
            +SnapshotDescription.table 
             
             
             private TableName
            -SnapshotDescription.table 
            +RegionCoprocessorRpcChannel.table 
             
             
             private TableName
            -HRegionLocator.tableName 
            +RawAsyncTableImpl.tableName 
             
             
             private TableName
            -ScannerCallableWithReplicas.tableName 
            +RegionServerCallable.tableName 
             
             
             protected TableName
            -ClientScanner.tableName 
            +RegionAdminServiceCallable.tableName 
             
             
             private TableName
            -AsyncClientScanner.tableName 
            +BufferedMutatorImpl.tableName 
             
             
             private TableName
            -AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.tableName 
            +AsyncProcessTask.tableName 
             
             
             private TableName
            -AsyncRpcRetryingCallerFactory.BatchCallerBuilder.tableName 
            +AsyncProcessTask.Builder.tableName 
             
             
             private TableName
            -RegionInfoBuilder.tableName 
            +AsyncRequestFutureImpl.tableName 
             
             
            -private TableName
            -RegionInfoBuilder.MutableRegionInfo.tableName 
            +protected TableName
            +TableBuilderBase.tableName 
             
             
             private TableName
            -RawAsyncTableImpl.tableName 
            +AsyncBatchRpcRetryingCaller.tableName 
             
             
             private TableName
            -RegionCoprocessorRpcChannelImpl.tableName 
            +RegionInfoBuilder.tableName 
             
             
             private TableName
            -AsyncTableRegionLocatorImpl.tableName 
            +RegionInfoBuilder.MutableRegionInfo.tableName 
             
             
            -protected TableName
            -RegionAdminServiceCallable.tableName 
            +private TableName
            +HTable.tableName 
             
             
             private TableName
            -HTable.tableName 
            +TableState.tableName 
             
             
            -private TableName
            -BufferedMutatorImpl.tableName 
            +protected TableName
            +RpcRetryingCallerWithReadReplicas.tableName 
             
             
            -private TableName
            -AsyncBatchRpcRetryingCaller.tableName 
            +protected TableName
            +AsyncTableBuilderBase.tableName 
             
             
             private TableName
            -BufferedMutatorParams.tableName 
            +AsyncSingleRequestRpcRetryingCaller.tableName 
             
             
             private TableName
            -HBaseAdmin.TableFuture.tableName 
            +ScannerCallableWithReplicas.tableName 
             
             
            -private TableName
            -AsyncRequestFutureImpl.tableName 
            +protected TableName
            +RawAsyncHBaseAdmin.TableProcedureBiConsumer.tableName 
             
             
             private TableName
            -AsyncProcessTask.tableName 
            +AsyncTableRegionLocatorImpl.tableName 
             
             
             private TableName
            -AsyncProcessTask.Builder.tableName 
            +HBaseAdmin.TableFuture.tableName 
             
             
            -protected TableName
            -RawAsyncHBaseAdmin.TableProcedureBiConsumer.tableName 
            +private TableName
            +RegionCoprocessorRpcChannelImpl.tableName 
             
             
            -private TableName
            -RegionServerCallable.tableName 
            +protected TableName
            +ClientScanner.tableName 
             
             
             private TableName
            -AsyncSingleRequestRpcRetryingCaller.tableName 
            +BufferedMutatorParams.tableName 
             
             
            -protected TableName
            -TableBuilderBase.tableName 
            +private TableName
            +AsyncClientScanner.tableName 
             
             
            -protected TableName
            -RpcRetryingCallerWithReadReplicas.tableName 
            +private TableName
            +AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.tableName 
             
             
            -protected TableName
            -AsyncTableBuilderBase.tableName 
            +private TableName
            +AsyncRpcRetryingCallerFactory.BatchCallerBuilder.tableName 
             
             
             private TableName
            -TableState.tableName 
            +HRegionLocator.tableName 
             
             
             
            @@ -2206,83 +2206,83 @@ service.
             
             
             TableName
            -AsyncTable.getName()
            -Gets the fully qualified table name instance of this 
            table.
            -
            +RawAsyncTableImpl.getName() 
             
             
             TableName
            -Table.getName()
            +RegionLocator.getName()
             Gets the fully qualified table name instance of this 
            table.
             
             
             
             TableName
            -HRegionLocator.getName() 
            +BufferedMutatorImpl.getName() 
             
             
             TableName
            -AsyncTableRegionLocator.getName()
            -Gets the fully qualified table name instance of the table 
            whose region we want to locate.
            +BufferedMutator.getName()
            +Gets the fully qualified table name instance of the table 
            that this BufferedMutator writes to.
             
             
             
             TableName
            -AsyncTableImpl.getName() 
            +HTable.getName() 
             
             
             TableName
            -RawAsyncTableImpl.getName() 
            +AsyncBufferedMutator.getName()
            +Gets the fully qualified table name instance of the table 
            that this
            + AsyncBufferedMutator writes to.
            +
             
             
             TableName
            -AsyncTableRegionLocatorImpl.getName() 
            +Table.getName()
            +Gets the fully qualified table name instance of this 
            table.
            +
             
             
             TableName
            -BufferedMutator.getName()
            -Gets the fully qualified table name instance of the table 
            that this BufferedMutator writes to.
            -
            +AsyncTableImpl.getName() 
             
             
             TableName
            -RegionLocator.getName()
            +AsyncTable.getName()
             Gets the fully qualified table name instance of this 
            table.
             
             
             
             TableName
            -AsyncBufferedMutatorImpl.getName() 
            +AsyncTableRegionLocatorImpl.getName() 
             
             
             TableName
            -HTable.getName() 
            +

            [27/51] [partial] hbase-site git commit: Published site at b7b86839250bf9b295ebc1948826f43a88736d6c.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
             
            b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
            index 2b5efbb..f0821d4 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
            @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
             
             
             Prev Class
            -Next Class
            +Next Class
             
             
             Frames
            @@ -74,7 +74,7 @@ var activeTableTab = "activeTableTab";
             
             
             Summary: 
            -Nested | 
            +Nested | 
             Field | 
             Constr | 
             Method
            @@ -126,7 +126,8 @@ var activeTableTab = "activeTableTab";
              @InterfaceStability.Evolving
             public class ReplicationSourceWALReader
             extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true";
             title="class or interface in java.lang">Thread
            -Reads and filters WAL entries, groups the filtered entries 
            into batches, and puts the batches onto a queue
            +Reads and filters WAL entries, groups the filtered entries 
            into batches, and puts the batches
            + onto a queue
             
             
             
            @@ -139,19 +140,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             
             
             Nested Class Summary
            -
            -Nested Classes 
            -
            -Modifier and Type
            -Class and Description
            -
            -
            -(package private) static class 
            -ReplicationSourceWALReader.WALEntryBatch
            -Holds a batch of WAL entries to replicate, along with some 
            statistics
            -
            -
            -
             
             
             
            @@ -182,7 +170,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             currentPosition 
             
             
            -protected https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true";
             title="class or interface in java.util.concurrent">BlockingQueue
            +protected https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true";
             title="class or interface in java.util.concurrent">BlockingQueue
             entryBatchQueue 
             
             
            @@ -222,18 +210,22 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             replicationBatchSizeCapacity 
             
             
            +private SerialReplicationChecker
            +serialReplicationChecker 
            +
            +
             private long
             sleepForRetries 
             
            -
            +
             private ReplicationSource
             source 
             
            -
            +
             private long
             totalBufferQuota 
             
            -
            +
             private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true";
             title="class or interface in 
            java.util.concurrent.atomic">AtomicLong
             totalBufferUsed 
             
            @@ -317,8 +309,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             
             
             protected void
            -handleEmptyWALEntryBatch(ReplicationSourceWALReader.WALEntryBatch batch,
            -
            org.apache.hadoop.fs.Path currentPath) 
            +handleEmptyWALEntryBatch(org.apache.hadoop.fs.Path currentPath) 
             
             
             private void
            @@ -329,7 +320,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             isReaderRunning() 
             
             
            -private ReplicationSourceWALReader.WALEntryBatch
            +private WALEntryBatch
             readWALEntries(WALEntryStream entryStream) 
             
             
            @@ -341,7 +332,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             setReaderRunning(boolean readerRunning) 
             
             
            -ReplicationSourceWALReader.WALEntryBatch
            +WALEntryBatch
             take()
             Retrieves the next batch of WAL entries from the queue, 
            waiting up to the specified time for a
              batch to become available
            @@ -349,9 +340,8 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             
             
             private void
            -updateBatchStats(ReplicationSourceWALReader.WALEntryBatch batch,
            +updateBatchStats(WALEntryBatch batch,
             WAL.Entry entry,
            -long entryPosition,
             long entrySize) 
             
             
            @@ -443,7 +433,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             
             
             entryBatchQueue
            -protected final https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true";
             title="class or interface in java.util.concurrent">BlockingQueue
             entryBatchQueue
            +protected final https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true";
             title="class or interface in java.util.concurrent">BlockingQueue entryBatchQueue
             
             
             
            @@ -521,12 +511,21 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             
             
             
            -
            +
             
             totalBufferQuota
             private long totalBufferQuota
             
             
            +
            +
            +
            +
            +
            +serialReplicationChecker
            +private final SerialReplicationChecker 
            serialReplicationChecker
            +
            +
             
             
             
            @@ -541,7 +540,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
             
             
             ReplicationSourceWALReader
            -pu

            [27/51] [partial] hbase-site git commit: Published site at 1d25b60831b8cc8f7ad5fd366f1867de5c20d2f3.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html
            --
            diff --git a/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html 
            b/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html
            index 6ea4d25..1900827 100644
            --- a/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html
            +++ b/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html
            @@ -117,7 +117,7 @@
             
             
             
            -static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             ServerName.EMPTY_SERVER_LIST 
             
             
            @@ -166,7 +166,7 @@
             
             
             static ServerName
            -ServerName.parseServerName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String str) 
            +ServerName.parseServerName(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String str) 
             
             
             static ServerName
            @@ -177,13 +177,13 @@
             
             
             static ServerName
            -ServerName.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String serverName)
            +ServerName.valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String serverName)
             Retrieve an instance of ServerName.
             
             
             
             static ServerName
            -ServerName.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String hostname,
            +ServerName.valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String hostname,
            int port,
            long startcode)
             Retrieve an instance of ServerName.
            @@ -191,7 +191,7 @@
             
             
             static ServerName
            -ServerName.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String hostAndPort,
            +ServerName.valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String hostAndPort,
            long startCode)
             Retrieve an instance of ServerName.
             
            @@ -206,17 +206,17 @@
             
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             ClusterMetrics.getBackupMasterNames() 
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             ClusterStatus.getBackupMasterNames()
             Deprecated. 
              
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             ClusterStatus.getBackupMasters()
             Deprecated. 
             As of release 2.0.0, this 
            will be removed in HBase 3.0.0
            @@ -225,27 +225,27 @@
             
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             ClusterMetrics.getDeadServerNames() 
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             ClusterStatus.getDeadServerNames()
             Deprecated. 
              
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
             title="class or interface in java.util">Map
            +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
             title="class or interface in java.util">Map
             ClusterMetrics.getLiveServerMetrics() 
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
             title="class or interface in java.util">Map
            +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
             title="class or interface in java.util">Map
             ClusterStatus.getLiveServerMetrics()
             Deprecated. 
              
             
             
            -http://docs.oracle.com/javase/8/

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html 
            b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
            index b419e6d..8e88e74 100644
            --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
            +++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
            @@ -2585,7 +2585,7 @@ implements 
             
             stopProcedureExecutor
            -private void stopProcedureExecutor()
            +private void stopProcedureExecutor()
             
             
             
            @@ -2594,7 +2594,7 @@ implements 
             
             stopChores
            -private void stopChores()
            +private void stopChores()
             
             
             
            @@ -2603,7 +2603,7 @@ implements 
             
             getRemoteInetAddress
            -http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
             title="class or interface in java.net">InetAddress getRemoteInetAddress(int port,
            +http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true";
             title="class or interface in java.net">InetAddress getRemoteInetAddress(int port,
              long serverStartCode)
               throws http://docs.oracle.com/javase/8/docs/api/java/net/UnknownHostException.html?is-external=true";
             title="class or interface in java.net">UnknownHostException
             
            @@ -2620,7 +2620,7 @@ implements 
             
             getMaxBalancingTime
            -private int getMaxBalancingTime()
            +private int getMaxBalancingTime()
             
             Returns:
             Maximum time we should run balancer for
            @@ -2633,7 +2633,7 @@ implements 
             
             getMaxRegionsInTransition
            -private int getMaxRegionsInTransition()
            +private int getMaxRegionsInTransition()
             
             Returns:
             Maximum number of regions in transition
            @@ -2646,7 +2646,7 @@ implements 
             
             balanceThrottling
            -private void balanceThrottling(long nextBalanceStartTime,
            +private void balanceThrottling(long nextBalanceStartTime,
            int maxRegionsInTransition,
            long cutoffTime)
             It first sleep to the next balance plan start time. 
            Meanwhile, throttling by the max
            @@ -2665,7 +2665,7 @@ implements 
             
             balance
            -public boolean balance()
            +public boolean balance()
             throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in java.io">IOException
             
             Throws:
            @@ -2679,7 +2679,7 @@ implements 
             
             balance
            -public boolean balance(boolean force)
            +public boolean balance(boolean force)
             throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in java.io">IOException
             
             Throws:
            @@ -2693,7 +2693,7 @@ implements 
             
             getRegionNormalizer
            -public RegionNormalizer getRegionNormalizer()
            +public RegionNormalizer getRegionNormalizer()
             
             Specified by:
             getRegionNormalizer in
             interface MasterServices
            @@ -2708,7 +2708,7 @@ implements 
             
             normalizeRegions
            -public boolean normalizeRegions()
            +public boolean normalizeRegions()
              throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in java.io">IOException
             Perform normalization of cluster (invoked by RegionNormalizerChore).
             
            @@ -2727,7 +2727,7 @@ implements 
             
             getClientIdAuditPrefix
            -public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String getClientIdAuditPrefix()
            +public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String getClientIdAuditPrefix()
             
             Specified by:
             getClientIdAuditPrefix in
             interface MasterServices
            @@ -2742,7 +2742,7 @@ implements 
             
             setCatalogJanitorEnabled
            -public void setCatalogJanitorEnabled(boolean b)
            +public void setCatalogJanitorEnabled(boolean b)
             Switch for the background CatalogJanitor thread.
              Used for testing.  The thread will continue to run.  It will just be a noop
              if disabled.
            @@ -2758,7 +2758,7 @@ implements 
             
             mergeRegions
            -public long mergeRegions(RegionInfo[] regionsToMerge,
            +public long mergeRegions(RegionInfo[] regionsToMerge,
              boolean forcible,
              long nonceGroup,
              long nonce)
            @@ -2786,7 +2786,7 @@ implements 
             
             splitRegion
            -public long splitRegion(RegionInfo regionInfo,
            +public long splitRegion(RegionInfo regionInfo,
             byte[] splitRow,
             long nonceGroup,
             long nonce)
            @@ -2814,7 +2814,7 @@ implements 
             
             move
            -public void move(byte[] encodedRegionName,
            +public void move(byte[] encodedRegionName,
              byte[] destServerName)
               throws HBaseIOException
             
            @@ -2829,7 +2829,7 @@ implements 
             
             createTable
            -public long createTable(TableDescriptor tableDescriptor,
            +public long createTable(TableDescriptor tableDescriptor,
             byte[][] splitKeys,
            

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
            index d143ef8..4583895 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
            @@ -258,7 +258,7 @@
             250
             251@Override
             252public long heapSize() {
            -253  long sum = HEAP_SIZE_OVERHEAD + 
            estimatedHeapSizeOf(cell);
            +253  long sum = HEAP_SIZE_OVERHEAD + 
            estimatedSizeOfCell(cell);
             254  if (this.tags != null) {
             255sum += 
            ClassSize.sizeOf(this.tags);
             256  }
            @@ -454,7 +454,7 @@
             446
             447@Override
             448public long heapSize() {
            -449  long sum = HEAP_SIZE_OVERHEAD + 
            estimatedHeapSizeOf(cell);
            +449  long sum = HEAP_SIZE_OVERHEAD + 
            estimatedSizeOfCell(cell);
             450  // this.tags is on heap byte[]
             451  if (this.tags != null) {
             452sum += 
            ClassSize.sizeOf(this.tags);
            @@ -2791,192 +2791,193 @@
             2783   * {@link HeapSize} we call {@link 
            HeapSize#heapSize()} so cell can give a correct value. In other
             2784   * cases we just consider the bytes 
            occupied by the cell components ie. row, CF, qualifier,
             2785   * timestamp, type, value and tags.
            -2786   * @param cell
            -2787   * @return estimate of the heap 
            space
            -2788   */
            -2789  public static long 
            estimatedHeapSizeOf(final Cell cell) {
            -2790if (cell instanceof HeapSize) {
            -2791  return ((HeapSize) 
            cell).heapSize();
            -2792}
            -2793// TODO: Add sizing of references 
            that hold the row, family, etc., arrays.
            -2794return 
            estimatedSerializedSizeOf(cell);
            -2795  }
            -2796
            -2797  /**
            -2798   * This method exists just to 
            encapsulate how we serialize keys. To be replaced by a factory that
            -2799   * we query to figure what the Cell 
            implementation is and then, what serialization engine to use
            -2800   * and further, how to serialize the 
            key for inclusion in hfile index. TODO.
            -2801   * @param cell
            -2802   * @return The key portion of the Cell 
            serialized in the old-school KeyValue way or null if passed
            -2803   * a null 
            cell
            -2804   */
            -2805  public static byte[] 
            getCellKeySerializedAsKeyValueKey(final Cell cell) {
            -2806if (cell == null) return null;
            -2807byte[] b = new 
            byte[KeyValueUtil.keyLength(cell)];
            -2808KeyValueUtil.appendKeyTo(cell, b, 
            0);
            -2809return b;
            -2810  }
            -2811
            -2812  /**
            -2813   * Create a Cell that is smaller than 
            all other possible Cells for the given Cell's row.
            -2814   * @param cell
            -2815   * @return First possible Cell on 
            passed Cell's row.
            -2816   */
            -2817  public static Cell 
            createFirstOnRow(final Cell cell) {
            -2818if (cell instanceof 
            ByteBufferExtendedCell) {
            -2819  return new 
            FirstOnRowByteBufferExtendedCell(
            -2820  ((ByteBufferExtendedCell) 
            cell).getRowByteBuffer(),
            -2821  ((ByteBufferExtendedCell) 
            cell).getRowPosition(), cell.getRowLength());
            -2822}
            -2823return new 
            FirstOnRowCell(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength());
            -2824  }
            -2825
            -2826  public static Cell 
            createFirstOnRow(final byte[] row, int roffset, short rlength) {
            -2827return new FirstOnRowCell(row, 
            roffset, rlength);
            -2828  }
            -2829
            -2830  public static Cell 
            createFirstOnRow(final byte[] row, final byte[] family, final byte[] col) {
            -2831return createFirstOnRow(row, 0, 
            (short) row.length, family, 0, (byte) family.length, col, 0,
            -2832col.length);
            -2833  }
            -2834
            -2835  public static Cell 
            createFirstOnRow(final byte[] row, int roffset, short rlength,
            -2836  final byte[] family, int foffset, 
            byte flength, final byte[] col, int coffset, int clength) {
            -2837return new FirstOnRowColCell(row, 
            roffset, rlength, family, foffset, flength, col, coffset,
            -2838clength);
            -2839  }
            -2840
            -2841  public static Cell 
            createFirstOnRow(final byte[] row) {
            -2842return createFirstOnRow(row, 0, 
            (short) row.length);
            -2843  }
            -2844
            -2845  public static Cell 
            createFirstOnRowFamily(Cell cell, byte[] fArray, int foff, int flen) {
            -2846if (cell instanceof 
            ByteBufferExtendedCell) {
            -2847  return new 
            FirstOnRowColByteBufferExtendedCell(
            -2848  ((ByteBufferExtendedCell) 
            cell).getRowByteBuffer(),
            -2849  ((ByteBufferExtendedCell) 
            cell).getRowPosition(), cell.getRowLength(),
            -2850  ByteBuffer.wrap(fArray), foff, 
            (byte) flen, HConstants.EMPTY_BYTE_BUFFER, 0, 0);
            -2851}
            -2852return new 
            FirstOnRowColCell(cell.getRowArray(), cell.getRowOffset(), 
            cell.getRowLength(),
            -2853fArray, foff, (byte) flen, 
            HConstants.EMPTY_BYTE_ARRAY, 0, 0);
            -2854  }
            -2855
            -2856  public static Cell 
            createFirstOnRowCol(final Cell cell) {
            -2857if (cell insta

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
            index cda7c78..93f650f 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
            @@ -269,7 +269,7 @@
             261this.rpcControllerFactory = 
            RpcControllerFactory.instantiate(conf);
             262this.rpcCallerFactory = 
            RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats);
             263this.backoffPolicy = 
            ClientBackoffPolicyFactory.create(conf);
            -264this.asyncProcess = new 
            AsyncProcess(this, conf, rpcCallerFactory, false, rpcControllerFactory);
            +264this.asyncProcess = new 
            AsyncProcess(this, conf, rpcCallerFactory, rpcControllerFactory);
             265if 
            (conf.getBoolean(CLIENT_SIDE_METRICS_ENABLED_KEY, false)) {
             266  this.metrics = new 
            MetricsConnection(this);
             267} else {
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
            index cda7c78..93f650f 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
            @@ -269,7 +269,7 @@
             261this.rpcControllerFactory = 
            RpcControllerFactory.instantiate(conf);
             262this.rpcCallerFactory = 
            RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats);
             263this.backoffPolicy = 
            ClientBackoffPolicyFactory.create(conf);
            -264this.asyncProcess = new 
            AsyncProcess(this, conf, rpcCallerFactory, false, rpcControllerFactory);
            +264this.asyncProcess = new 
            AsyncProcess(this, conf, rpcCallerFactory, rpcControllerFactory);
             265if 
            (conf.getBoolean(CLIENT_SIDE_METRICS_ENABLED_KEY, false)) {
             266  this.metrics = new 
            MetricsConnection(this);
             267} else {
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
            index cda7c78..93f650f 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
            @@ -269,7 +269,7 @@
             261this.rpcControllerFactory = 
            RpcControllerFactory.instantiate(conf);
             262this.rpcCallerFactory = 
            RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats);
             263this.backoffPolicy = 
            ClientBackoffPolicyFactory.create(conf);
            -264this.asyncProcess = new 
            AsyncProcess(this, conf, rpcCallerFactory, false, rpcControllerFactory);
            +264this.asyncProcess = new 
            AsyncProcess(this, conf, rpcCallerFactory, rpcControllerFactory);
             265if 
            (conf.getBoolean(CLIENT_SIDE_METRICS_ENABLED_KEY, false)) {
             266  this.metrics = new 
            MetricsConnection(this);
             267} else {
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
            index cda7c78..93f650f 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
            @@ -269,7 +269,7 @@
             261this.rpcControllerFactory = 
            RpcControllerFactory.instantiate(conf);
             262this.rpcCallerFactory = 
            RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats);
             263this.backoffPolicy = 
            ClientBackoffPolicyFactory.create(con

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
            index 90f52b0..2ac1b78 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
            @@ -106,7 +106,7 @@
             
             
             private RegionLocateType
            -AsyncSingleRequestRpcRetryingCaller.locateType 
            +AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.locateType 
             
             
             RegionLocateType
            @@ -114,7 +114,7 @@
             
             
             private RegionLocateType
            -AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.locateType 
            +AsyncSingleRequestRpcRetryingCaller.locateType 
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
            index e062eb5..fbe0658 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
            @@ -230,14 +230,14 @@ service.
             
             
             private RegionLocator
            -HFileOutputFormat2.TableInfo.regionLocator 
            -
            -
            -private RegionLocator
             TableInputFormatBase.regionLocator
             The RegionLocator of the 
            table.
             
             
            +
            +private RegionLocator
            +HFileOutputFormat2.TableInfo.regionLocator 
            +
             
             
             
            @@ -248,15 +248,15 @@ service.
             
             
             
            -RegionLocator
            -HFileOutputFormat2.TableInfo.getRegionLocator() 
            -
            -
             protected RegionLocator
             TableInputFormatBase.getRegionLocator()
             Allows subclasses to get the RegionLocator.
             
             
            +
            +RegionLocator
            +HFileOutputFormat2.TableInfo.getRegionLocator() 
            +
             
             
             
            
            
            

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
            index ded834e..9c4acf5 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
            @@ -330,8 +330,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            -AsyncHBaseAdmin.addColumnFamily(TableName tableName,
            -   ColumnFamilyDescriptor columnFamily) 
            +AsyncAdmin.addColumnFamily(TableName tableName,
            +   ColumnFamilyDescriptor columnFamily)
            +Add a column family to an existing table.
            +
             
             
             void
            @@ -341,20 +343,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -void
            -HBaseAdmin.addColumnFamily(TableName tableName,
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            +RawAsyncHBaseAdmin.addColumnFamily(TableName tableName,
            ColumnFamilyDescriptor columnFamily) 
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            -AsyncAdmin.addColumnFamily(TableName tableName,
            -   ColumnFamilyDescriptor columnFamily)
            -Add a column family to an existing table.
            -
            +void
            +HBaseAdmin.addColumnFamily(TableName tableName,
            +   ColumnFamilyDescriptor columnFamily) 
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            -RawAsyncHBaseAdmin.addColumnFamily(TableName tableName,
            +AsyncHBaseAdmin.addColumnFamily(TableName tableName,
            ColumnFamilyDescriptor columnFamily) 
             
             
            @@ -396,8 +396,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            -AsyncHBaseAdmin.modifyColumnFamily(TableName tableName,
            -  ColumnFamilyDescriptor columnFamily) 
            +AsyncAdmin.modifyColumnFamily(TableName tableName,
            +  ColumnFamilyDescriptor columnFamily)
            +Modify an existing column family on a table.
            +
             
             
             void
            @@ -407,20 +409,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -void
            -HBaseAdmin.modifyColumnFamily(TableName tableName,
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            +RawAsyncHBaseAdmin.modifyColumnFamily(TableName tableName,
               ColumnFamilyDescriptor columnFamily) 
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            -AsyncAdmin.modifyColumnFamily(TableName tableName,
            -  ColumnFamilyDescriptor columnFamily)
            -Modify an existing column family on a table.
            -
            +void
            +HBaseAdmin.modifyColumnFamily(TableName tableName,
            +  ColumnFamilyDescriptor columnFamily) 
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFuture[27/51] [partial] hbase-site git commit: Published site at .
            
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
            index 53d69df..597d1ba 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
            @@ -355,24 +355,24 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             void
            -BufferedMutatorImpl.mutate(Mutation m) 
            -
            -
            -void
             BufferedMutator.mutate(Mutation mutation)
             Sends a Mutation to 
            the table.
             
             
            +
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            +AsyncBufferedMutatorImpl.mutate(Mutation mutation) 
            +
             
            +void
            +BufferedMutatorImpl.mutate(Mutation m) 
            +
            +
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
             AsyncBufferedMutator.mutate(Mutation mutation)
             Sends a Mutation to 
            the table.
             
             
            -
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            -AsyncBufferedMutatorImpl.mutate(Mutation mutation) 
            -
             
             
             
            @@ -390,24 +390,24 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             void
            -BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List ms) 
            -
            -
            -void
             BufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations)
             Send some Mutations to 
            the table.
             
             
            +
            +http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">ListCompletableFutureVoid>>
            +AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations) 
            +
             
            +void
            +BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List ms) 
            +
            +
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">ListCompletableFutureVoid>>
             AsyncBufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations)
             Send some Mutations to 
            the table.
             
             
            -
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">ListCompletableFutureVoid>>
            -AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations) 
            -
             
             static RowMutations
             RowMutations.of(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations)
            @@ -543,15 +543,15 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             MutationSerialization.getDeserializer(http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true";
             title="class or interface in java.lang">Class
            

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            index 1d50582..7eb7661 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            @@ -104,14 +104,14 @@
             
             
             void
            -HMaster.checkTableModifiable(TableName tableName) 
            -
            -
            -void
             MasterServices.checkTableModifiable(TableName tableName)
             Check table is modifiable; i.e.
             
             
            +
            +void
            +HMaster.checkTableModifiable(TableName tableName) 
            +
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            index 5eebc7c..b1a475b 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            @@ -144,6 +144,19 @@
             
             
             Uses of TableNotFoundException in org.apache.hadoop.hbase.master
            +
            +Subclasses of TableNotFoundException in org.apache.hadoop.hbase.master 
            +
            +Modifier and Type
            +Class and Description
            +
            +
            +
            +static class 
            +TableStateManager.TableStateNotFoundException 
            +
            +
            +
             
             Methods in org.apache.hadoop.hbase.master
             that throw TableNotFoundException 
             
            @@ -157,14 +170,14 @@
             
             
             void
            -HMaster.checkTableModifiable(TableName tableName) 
            -
            -
            -void
             MasterServices.checkTableModifiable(TableName tableName)
             Check table is modifiable; i.e.
             
             
            +
            +void
            +HMaster.checkTableModifiable(TableName tableName) 
            +
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            index 594e74b..90f6967 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            @@ -243,6 +243,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +PrivateCellUtil.getTags(Cell cell) 
            +
            +
            +static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
             CellUtil.getTags(Cell cell)
             Deprecated. 
             As of release 2.0.0, this 
            will be removed in HBase 3.0.0.
            @@ -250,10 +254,6 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -
            -static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            -PrivateCellUtil.getTags(Cell cell) 
            -
             
             static http://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true";
             title="class or interface in java.util">Iterator
             CellUtil.tagsIterator(byte[] tags,
            @@ -395,6 +395,11 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             static Cell
            +PrivateCellUtil.createCell(Cell cell,
            +  http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List tags) 
            +
            +
            +static Cell
             CellUtil.createCell(Cell cell,
               http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List tags)
             Deprecated. 
            @@ -402,11 +407,6 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -
            -static Cell
            -PrivateCellUtil.createCell(Cell cell,
            -  http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List tags) 
            -
             
             static byte[]
             TagUtil.fromList(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List tags)
            @@ -415,16 +415,16 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -RawCellBuilder
            -RawCellBuilder.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List tags) 
            +ExtendedCellBuilder
            +ExtendedCellBuilderImpl.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List tags) 
             
             
            -ExtendedCellBuilder
            -ExtendedCellBuilder.setTags(http://doc

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
            index 53d69df..597d1ba 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
            @@ -355,24 +355,24 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             void
            -BufferedMutatorImpl.mutate(Mutation m) 
            -
            -
            -void
             BufferedMutator.mutate(Mutation mutation)
             Sends a Mutation to 
            the table.
             
             
            +
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            +AsyncBufferedMutatorImpl.mutate(Mutation mutation) 
            +
             
            +void
            +BufferedMutatorImpl.mutate(Mutation m) 
            +
            +
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
             AsyncBufferedMutator.mutate(Mutation mutation)
             Sends a Mutation to 
            the table.
             
             
            -
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureVoid>
            -AsyncBufferedMutatorImpl.mutate(Mutation mutation) 
            -
             
             
             
            @@ -390,24 +390,24 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             void
            -BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List ms) 
            -
            -
            -void
             BufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations)
             Send some Mutations to 
            the table.
             
             
            +
            +http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">ListCompletableFutureVoid>>
            +AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations) 
            +
             
            +void
            +BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List ms) 
            +
            +
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">ListCompletableFutureVoid>>
             AsyncBufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations)
             Send some Mutations to 
            the table.
             
             
            -
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">ListCompletableFutureVoid>>
            -AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations) 
            -
             
             static RowMutations
             RowMutations.of(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List mutations)
            @@ -543,15 +543,15 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             MutationSerialization.getDeserializer(http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true";
             title="class or interface in java.lang">Class
            

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            index 232a8b4..aeaf9fe 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            @@ -495,7 +495,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             private static HRegionLocation
            -AsyncMetaTableAccessor.getRegionLocation(Result r,
            +MetaTableAccessor.getRegionLocation(Result r,
              RegionInfo regionInfo,
              int replicaId)
             Returns the HRegionLocation parsed from the given meta row 
            Result
            @@ -504,7 +504,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             private static HRegionLocation
            -MetaTableAccessor.getRegionLocation(Result r,
            +AsyncMetaTableAccessor.getRegionLocation(Result r,
              RegionInfo regionInfo,
              int replicaId)
             Returns the HRegionLocation parsed from the given meta row 
            Result
            @@ -944,7 +944,9 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            -AsyncHBaseAdmin.getRegions(ServerName serverName) 
            +AsyncAdmin.getRegions(ServerName serverName)
            +Get all the online regions on a region server.
            +
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            @@ -953,22 +955,22 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            -HBaseAdmin.getRegions(ServerName sn) 
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            +RawAsyncHBaseAdmin.getRegions(ServerName serverName) 
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            -AsyncAdmin.getRegions(ServerName serverName)
            -Get all the online regions on a region server.
            -
            +http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            +HBaseAdmin.getRegions(ServerName sn) 
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            -RawAsyncHBaseAdmin.getRegions(ServerName serverName) 
            +AsyncHBaseAdmin.getRegions(ServerName serverName) 
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            -AsyncHBaseAdmin.getRegions(TableName tableName) 
            +AsyncAdmin.getRegions(TableName tableName)
            +Get the regions of a given table.
            +
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            @@ -977,18 +979,16 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List
            -HBaseAdmin.getRegions(TableName tableName) 
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            +RawAsyncHBaseAdmin.getRegions(TableName tableName) 
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=tr

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
            index ad601c4..53e455f 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
            @@ -1117,1183 +1117,1186 @@
             1109  @Nullable
             1110  public static TableState 
            getTableState(Connection conn, TableName tableName)
               throws IOException {
            -1112Table metaHTable = 
            getMetaHTable(conn);
            -1113Get get = new 
            Get(tableName.getName()).addColumn(getTableFamily(), getTableStateColumn());
            -1114long time = 
            EnvironmentEdgeManager.currentTime();
            -1115get.setTimeRange(0, time);
            -1116Result result =
            -1117metaHTable.get(get);
            -1118return getTableState(result);
            -1119  }
            -1120
            -1121  /**
            -1122   * Fetch table states from META 
            table
            -1123   * @param conn connection to use
            -1124   * @return map {tableName -> 
            state}
            -1125   * @throws IOException
            -1126   */
            -1127  public static Map getTableStates(Connection conn)
            -1128  throws IOException {
            -1129final Map states = new LinkedHashMap<>();
            -1130Visitor collector = new Visitor() 
            {
            -1131  @Override
            -1132  public boolean visit(Result r) 
            throws IOException {
            -1133TableState state = 
            getTableState(r);
            -1134if (state != null)
            -1135  
            states.put(state.getTableName(), state);
            -1136return true;
            -1137  }
            -1138};
            -1139fullScanTables(conn, collector);
            -1140return states;
            -1141  }
            -1142
            -1143  /**
            -1144   * Updates state in META
            -1145   * @param conn connection to use
            -1146   * @param tableName table to look 
            for
            -1147   * @throws IOException
            -1148   */
            -1149  public static void 
            updateTableState(Connection conn, TableName tableName,
            -1150  TableState.State actual) throws 
            IOException {
            -1151updateTableState(conn, new 
            TableState(tableName, actual));
            -1152  }
            -1153
            -1154  /**
            -1155   * Decode table state from META 
            Result.
            -1156   * Should contain cell from 
            HConstants.TABLE_FAMILY
            -1157   * @param r result
            -1158   * @return null if not found
            -1159   * @throws IOException
            -1160   */
            -1161  @Nullable
            -1162  public static TableState 
            getTableState(Result r)
            -1163  throws IOException {
            -1164Cell cell = 
            r.getColumnLatestCell(getTableFamily(), getTableStateColumn());
            -1165if (cell == null) return null;
            -1166try {
            -1167  return 
            TableState.parseFrom(TableName.valueOf(r.getRow()),
            -1168  
            Arrays.copyOfRange(cell.getValueArray(),
            -1169  cell.getValueOffset(), 
            cell.getValueOffset() + cell.getValueLength()));
            -1170} catch (DeserializationException e) 
            {
            -1171  throw new IOException(e);
            -1172}
            -1173
            -1174  }
            -1175
            -1176  /**
            -1177   * Implementations 'visit' a catalog 
            table row.
            -1178   */
            -1179  public interface Visitor {
            -1180/**
            -1181 * Visit the catalog table row.
            -1182 * @param r A row from catalog 
            table
            -1183 * @return True if we are to proceed 
            scanning the table, else false if
            -1184 * we are to stop now.
            -1185 */
            -1186boolean visit(final Result r) throws 
            IOException;
            -1187  }
            -1188
            -1189  /**
            -1190   * Implementations 'visit' a catalog 
            table row but with close() at the end.
            -1191   */
            -1192  public interface CloseableVisitor 
            extends Visitor, Closeable {
            -1193  }
            -1194
            -1195  /**
            -1196   * A {@link Visitor} that collects 
            content out of passed {@link Result}.
            -1197   */
            -1198  static abstract class 
            CollectingVisitor implements Visitor {
            -1199final List results = new 
            ArrayList<>();
            -1200@Override
            -1201public boolean visit(Result r) 
            throws IOException {
            -1202  if (r ==  null || r.isEmpty()) 
            return true;
            -1203  add(r);
            -1204  return true;
            -1205}
            -1206
            -1207abstract void add(Result r);
            -1208
            -1209/**
            -1210 * @return Collected results; wait 
            till visits complete to collect all
            -1211 * possible results
            -1212 */
            -1213List getResults() {
            -1214  return this.results;
            -1215}
            -1216  }
            -1217
            -1218  /**
            -1219   * Collects all returned.
            -1220   */
            -1221  static class CollectAllVisitor extends 
            CollectingVisitor {
            -1222@Override
            -1223void add(Result r) {
            -1224  this.results.add(r);
            -1225}
            -1226  }
            -1227
            -1228  /**
            -1229   * A Visitor that skips offline 
            regions and split parents
            -1230   */
            -1231  public static abstract class 
            DefaultVisitorBase implements Visitor {
            -1232
            -1233public DefaultVisitorBase() {
            -1234  super();
            -1235}
            -1236
            -1237public abstract boolean 
            visitInternal(Result rowResult) throws IOException;
            -1238
            

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html 
            b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
            index ea234c2..34c375d 100644
            --- a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
            +++ b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
            @@ -567,7 +567,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             warningMap
            -private static final http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
             title="class or interface in 
            java.util">MapBoolean> warningMap
            +private static final http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
             title="class or interface in 
            java.util">MapBoolean> warningMap
             
             
             
            @@ -972,7 +972,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             isValidWALRootDir
            -private static boolean isValidWALRootDir(org.apache.hadoop.fs.Path walDir,
            +private static boolean isValidWALRootDir(org.apache.hadoop.fs.Path walDir,
              
            org.apache.hadoop.conf.Configuration c)
               throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in java.io">IOException
             
            @@ -987,7 +987,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             getTableDir
            -public static org.apache.hadoop.fs.Path getTableDir(org.apache.hadoop.fs.Path rootdir,
            +public static org.apache.hadoop.fs.Path getTableDir(org.apache.hadoop.fs.Path rootdir,
             TableName tableName)
             Returns the Path object representing the table 
            directory under
              path rootdir
            @@ -1006,7 +1006,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             getTableName
            -public static TableName getTableName(org.apache.hadoop.fs.Path tablePath)
            +public static TableName getTableName(org.apache.hadoop.fs.Path tablePath)
             Returns the TableName object representing
              the table directory under
              path rootdir
            @@ -1024,7 +1024,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             getNamespaceDir
            -public static org.apache.hadoop.fs.Path getNamespaceDir(org.apache.hadoop.fs.Path rootdir,
            +public static org.apache.hadoop.fs.Path getNamespaceDir(org.apache.hadoop.fs.Path rootdir,
             http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String namespace)
             Returns the Path object representing
              the namespace directory under path rootdir
            @@ -1043,7 +1043,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             setStoragePolicy
            -public static void setStoragePolicy(org.apache.hadoop.fs.FileSystem fs,
            +public static void setStoragePolicy(org.apache.hadoop.fs.FileSystem fs,
             
            org.apache.hadoop.conf.Configuration conf,
             org.apache.hadoop.fs.Path path,
             http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String policyKey,
            @@ -1074,7 +1074,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             setStoragePolicy
            -public static void setStoragePolicy(org.apache.hadoop.fs.FileSystem fs,
            +public static void setStoragePolicy(org.apache.hadoop.fs.FileSystem fs,
             org.apache.hadoop.fs.Path path,
             http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String storagePolicy)
             Sets storage policy for given path.
            @@ -1101,7 +1101,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             invokeSetStoragePolicy
            -private static void invokeSetStoragePolicy(org.apache.hadoop.fs.FileSystem fs,
            +private static void invokeSetStoragePolicy(org.apache.hadoop.fs.FileSystem fs,
            org.apache.hadoop.fs.Path path,
            http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String storagePolicy)
             
            @@ -1112,7 +1112,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             isHDFS
            -public static boolean isHDFS(org.apache.hadoop.conf.Config

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
            --
            diff --git 
            a/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
             
            b/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
            index 633812a..d239905 100644
            --- 
            a/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
            +++ 
            b/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
            @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -public class TestServerSideScanMetricsFromClientSide
            +public class TestServerSideScanMetricsFromClientSide
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object
             
             
            @@ -134,64 +134,68 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             CELL_HEAP_SIZE 
             
             
            +static HBaseClassTestRule
            +CLASS_RULE 
            +
            +
             private static byte[][]
             FAMILIES 
             
            -
            +
             private static byte[]
             FAMILY 
             
            -
            +
             private static int
             NUM_COLS 
             
            -
            +
             private static int
             NUM_FAMILIES 
             
            -
            +
             private static int
             NUM_QUALIFIERS 
             
            -
            +
             private static int
             NUM_ROWS 
             
            -
            +
             private static byte[]
             QUALIFIER 
             
            -
            +
             private static byte[][]
             QUALIFIERS 
             
            -
            +
             private static byte[]
             ROW 
             
            -
            +
             private static byte[][]
             ROWS 
             
            -
            +
             private static 
            org.apache.hadoop.hbase.client.Table
             TABLE 
             
            -
            +
             private static 
            org.apache.hadoop.hbase.TableName
             TABLE_NAME
             Table configuration
             
             
            -
            +
             private static HBaseTestingUtility
             TEST_UTIL 
             
            -
            +
             private static byte[]
             VALUE 
             
            -
            +
             private static int
             VALUE_SIZE 
             
            @@ -316,13 +320,22 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             Field Detail
            +
            +
            +
            +
            +
            +CLASS_RULE
            +public static final HBaseClassTestRule CLASS_RULE
            +
            +
             
             
             
             
             
             TEST_UTIL
            -private static final HBaseTestingUtility TEST_UTIL
            +private static final HBaseTestingUtility TEST_UTIL
             
             
             
            @@ -331,7 +344,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             TABLE
            -private static org.apache.hadoop.hbase.client.Table TABLE
            +private static org.apache.hadoop.hbase.client.Table TABLE
             
             
             
            @@ -340,7 +353,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             TABLE_NAME
            -private static org.apache.hadoop.hbase.TableName TABLE_NAME
            +private static org.apache.hadoop.hbase.TableName TABLE_NAME
             Table configuration
             
             
            @@ -350,7 +363,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             NUM_ROWS
            -private static int NUM_ROWS
            +private static int NUM_ROWS
             
             
             
            @@ -359,7 +372,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             ROW
            -private static byte[] ROW
            +private static byte[] ROW
             
             
             
            @@ -368,7 +381,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             ROWS
            -private static byte[][] ROWS
            +private static byte[][] ROWS
             
             
             
            @@ -377,7 +390,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             NUM_FAMILIES
            -private static int NUM_FAMILIES
            +private static int NUM_FAMILIES
             
             
             
            @@ -386,7 +399,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             FAMILY
            -private static byte[] FAMILY
            +private static byte[] FAMILY
             
             
             
            @@ -395,7 +408,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             FAMILIES
            -private static byte[][] FAMILIES
            +private static byte[][] FAMILIES
             
             
             
            @@ -404,7 +417,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             NUM_QUALIFIERS
            -private static int NUM_QUALIFIERS
            +private static int NUM_QUALIFIERS
             
             
             
            @@ -413,7 +426,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             QUALIFIER
            -private static byte[] QUALIFIER
            +private static byte[] QUALIFIER
             
             
             
            @@ -422,7 +435,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             QUALIFIERS
            -private static byte[][] QUALIFIERS
            +private static byte[][] QUALIFIERS
             
             
             
            @@ -431,7 +444,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             VALUE_SIZE
            -private static int VALUE_SIZE
            +private static int VALUE_SIZE
             
             
             
            @@ -440,7 +453,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             VALUE
            -private static byte[] VALUE
            +private static byte[] VALUE
             
             
             
            @@ -449,7 +462,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             NUM_COLS
            -private static int NUM_COLS
            +private static int NUM_COLS
             
             
             
            @@ -458,7 +471,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             CELL_HEAP_SIZE
            -private static long CELL_HEAP_SIZE
            +private static long CELL_HEAP_SIZE
             
             
             
            @@ -475,7 +488,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             TestServerSideScanMetricsFromClientSide

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
            index eb9e252..667152a 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
            @@ -28,22 +28,22 @@
             020
             021import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
             022import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
            -023import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
            -024import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
            -025import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
            -026import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
            -027import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
            -028import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
            -029import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
            -030import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
            -031import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
            -032import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
            -033import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
            -034import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
            -035import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
            -036import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
            -037import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
            -038import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
            +023import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
            +024import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
            +025import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
            +026import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
            +027import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
            +028import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
            +029import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
            +030import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
            +031import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
            +032import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
            +033import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
            +034import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
            +035import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
            +036import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
            +037import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
            +038import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
             039
             040import java.io.IOException;
             041import java.net.URI;
            @@ -70,194 +70,194 @@
             062import 
            org.apache.hadoop.hbase.backup.util.BackupUtils;
             063import 
            org.apache.hadoop.hbase.client.Connection;
             064import 
            org.apache.hadoop.hbase.client.ConnectionFactory;
            -065import 
            org.apache.hbase.thirdparty.com.google.common.collect.Lists;
            -066import 
            org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
            -067import 
            org.apache.yetus.audience.InterfaceAudience;
            -068
            -069/**
            -070 * General backup commands, options and 
            usage messages
            -071 */
            -072
            +065import 
            org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
            +066import 
            org.apache.yetus.audience.InterfaceAudience;
            +067
            +068import 
            org.apache.hbase.thirdparty.com.google.common.collect.Lists;
            +069
            +070/**
            +071 * General backup commands, options and 
            usage messages
            +072 */
             073@InterfaceAudience.Private
             074public final class BackupCommands {
            -075
            -076  public final static String 
            INCORRECT_USAGE = "Incorrect usage";
            -077
            -078  public final static String 
            TOP_LEVEL_NOT_ALLOWED =
            -079  "Top level (root) folder is not 
            allowed to be a backup destination";
            -080
            -081  public static final String USAGE = 
            "Usage: hbase bac

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
            index 0164ad3..fd16dee 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
            @@ -76,64 +76,54 @@
             068  }
             069
             070  @Override
            -071  void setTimeProgress(long timeProgress) 
            {
            +071  void setProgress(int batchProgress, 
            long sizeProgress, long heapSizeProgress) {
             072// Do nothing. NoLimitScannerContext 
            instances are immutable post-construction
             073  }
             074
             075  @Override
            -076  void updateTimeProgress() {
            +076  void clearProgress() {
             077// Do nothing. NoLimitScannerContext 
            instances are immutable post-construction
             078  }
             079
             080  @Override
            -081  void setProgress(int batchProgress, 
            long sizeProgress, long heapSizeProgress, long timeProgress) {
            +081  void setSizeLimitScope(LimitScope 
            scope) {
             082// Do nothing. NoLimitScannerContext 
            instances are immutable post-construction
             083  }
             084
             085  @Override
            -086  void clearProgress() {
            +086  void setTimeLimitScope(LimitScope 
            scope) {
             087// Do nothing. NoLimitScannerContext 
            instances are immutable post-construction
             088  }
             089
             090  @Override
            -091  void setSizeLimitScope(LimitScope 
            scope) {
            +091  NextState setScannerState(NextState 
            state) {
             092// Do nothing. NoLimitScannerContext 
            instances are immutable post-construction
            -093  }
            -094
            -095  @Override
            -096  void setTimeLimitScope(LimitScope 
            scope) {
            -097// Do nothing. NoLimitScannerContext 
            instances are immutable post-construction
            -098  }
            -099
            -100  @Override
            -101  NextState setScannerState(NextState 
            state) {
            -102// Do nothing. NoLimitScannerContext 
            instances are immutable post-construction
            -103return state;
            -104  }
            -105
            -106  @Override
            -107  boolean checkBatchLimit(LimitScope 
            checkerScope) {
            -108// No limits can be specified, thus 
            return false to indicate no limit has been reached.
            -109return false;
            -110  }
            -111
            -112  @Override
            -113  boolean checkSizeLimit(LimitScope 
            checkerScope) {
            -114// No limits can be specified, thus 
            return false to indicate no limit has been reached.
            -115return false;
            -116  }
            -117
            -118  @Override
            -119  boolean checkTimeLimit(LimitScope 
            checkerScope) {
            -120// No limits can be specified, thus 
            return false to indicate no limit has been reached.
            -121return false;
            -122  }
            -123
            -124  @Override
            -125  boolean checkAnyLimitReached(LimitScope 
            checkerScope) {
            -126return false;
            -127  }
            -128}
            +093return state;
            +094  }
            +095
            +096  @Override
            +097  boolean checkBatchLimit(LimitScope 
            checkerScope) {
            +098// No limits can be specified, thus 
            return false to indicate no limit has been reached.
            +099return false;
            +100  }
            +101
            +102  @Override
            +103  boolean checkSizeLimit(LimitScope 
            checkerScope) {
            +104// No limits can be specified, thus 
            return false to indicate no limit has been reached.
            +105return false;
            +106  }
            +107
            +108  @Override
            +109  boolean checkTimeLimit(LimitScope 
            checkerScope) {
            +110// No limits can be specified, thus 
            return false to indicate no limit has been reached.
            +111return false;
            +112  }
            +113
            +114  @Override
            +115  boolean checkAnyLimitReached(LimitScope 
            checkerScope) {
            +116return false;
            +117  }
            +118}
             
             
             
            
            
            

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html 
            b/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html
            index f8c49cb..e3762b3 100644
            --- a/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html
            +++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html
            @@ -1265,7 +1265,7 @@ implements 
             
             CACHE_FIXED_OVERHEAD
            -public static final long CACHE_FIXED_OVERHEAD
            +public static final long CACHE_FIXED_OVERHEAD
             
             
             
            @@ -1434,7 +1434,7 @@ implements 
             
             cacheBlock
            -public void cacheBlock(BlockCacheKey cacheKey,
            +public void cacheBlock(BlockCacheKey cacheKey,
            Cacheable buf)
             Cache the block with the specified name and buffer.
              
            @@ -1453,7 +1453,7 @@ implements 
             
             updateSizeMetrics
            -private long updateSizeMetrics(LruCachedBlock cb,
            +private long updateSizeMetrics(LruCachedBlock cb,
            boolean evict)
             Helper function that updates the local size counter and 
            also updates any
              per-cf or per-blocktype metrics it can discern from given
            @@ -1466,7 +1466,7 @@ implements 
             
             getBlock
            -public Cacheable getBlock(BlockCacheKey cacheKey,
            +public Cacheable getBlock(BlockCacheKey cacheKey,
               boolean caching,
               boolean repeat,
               boolean updateCacheMetrics)
            @@ -1492,7 +1492,7 @@ implements 
             
             containsBlock
            -public boolean containsBlock(BlockCacheKey cacheKey)
            +public boolean containsBlock(BlockCacheKey cacheKey)
             Whether the cache contains block with specified 
            cacheKey
             
             Returns:
            @@ -1506,7 +1506,7 @@ implements 
             
             evictBlock
            -public boolean evictBlock(BlockCacheKey cacheKey)
            +public boolean evictBlock(BlockCacheKey cacheKey)
             Description copied from 
            interface: BlockCache
             Evict block from cache.
             
            @@ -1525,7 +1525,7 @@ implements 
             
             evictBlocksByHfileName
            -public int evictBlocksByHfileName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String hfileName)
            +public int evictBlocksByHfileName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String hfileName)
             Evicts all blocks for a specific HFile. This is an
              expensive operation implemented as a linear-time search through all blocks
              in the cache. Ideally this should be a search in a log-access-time map.
            @@ -1546,7 +1546,7 @@ implements 
             
             evictBlock
            -protected long evictBlock(LruCachedBlock block,
            +protected long evictBlock(LruCachedBlock block,
               boolean evictedByEvictionProcess)
             Evict the block, and it will be cached by the victim 
            handler if exists &&
              block may be read again later
            @@ -1565,7 +1565,7 @@ implements 
             
             runEviction
            -private void runEviction()
            +private void runEviction()
             Multi-threaded call to run the eviction process.
             
             
            @@ -1575,7 +1575,7 @@ implements 
             
             isEvictionInProgress
            -boolean isEvictionInProgress()
            +boolean isEvictionInProgress()
             
             
             
            @@ -1584,7 +1584,7 @@ implements 
             
             getOverhead
            -long getOverhead()
            +long getOverhead()
             
             
             
            @@ -1593,7 +1593,7 @@ implements 
             
             evict
            -void evict()
            +void evict()
             Eviction method.
             
             
            @@ -1603,7 +1603,7 @@ implements 
             
             toString
            -public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String toString()
            +public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String toString()
             
             Overrides:
             http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--";
             title="class or interface in java.lang">toString in 
            class http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object
            @@ -1616,7 +1616,7 @@ implements 
             
             getMaxSize
            -public long getMaxSize()
            +public long getMaxSize()
             Get the maximum size of this cache.
             
             Specified by:
            @@ -1632,7 +1632,7 @@ implements 
             
             getCurrentSize
            -public long getCurrentSize()
            +public long getCurrentSize()
             Description copied from 
            interface: BlockCache
             Returns the occupied size of the block cache, in 
            bytes.
             
            @@ -1649,7 +1649,7 @@ implements 
             
             getCurrentDataSize
            -public long getCurrentDataSize()
            +public long getCurrentDataSize()
             Description copied from 
            interface: BlockCache
             Returns the occupied size of data blocks, in bytes.
             
            @@ -1666,7 +1666,7 @@ implements 
             
             getFreeSize
            -public long getFreeSize()
            +public long getFreeSize()
             Description copied from 
            interface: BlockCache
             Returns the free size of the block cache, in bytes.
             
            @@ -1683,7 +1683,7 @@ implements 
             
             size
            -public long size()
            +public long size()
             Description copied from 
            interface: BlockCache
             Returns the total size of

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
            index e05510e..2e114d8 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
            @@ -54,891 +54,884 @@
             046import 
            org.apache.hadoop.hbase.HTableDescriptor;
             047import 
            org.apache.hadoop.hbase.MetaTableAccessor;
             048import 
            org.apache.hadoop.hbase.MetaTableAccessor.DefaultVisitorBase;
            -049import 
            org.apache.hadoop.hbase.NamespaceDescriptor;
            -050import 
            org.apache.hadoop.hbase.ServerName;
            -051import 
            org.apache.hadoop.hbase.TableName;
            -052import 
            org.apache.hadoop.hbase.client.ClusterConnection;
            -053import 
            org.apache.hadoop.hbase.client.Delete;
            -054import 
            org.apache.hadoop.hbase.client.Get;
            -055import 
            org.apache.hadoop.hbase.client.Mutation;
            -056import 
            org.apache.hadoop.hbase.client.Put;
            -057import 
            org.apache.hadoop.hbase.client.RegionInfo;
            -058import 
            org.apache.hadoop.hbase.client.Result;
            -059import 
            org.apache.hadoop.hbase.client.Scan;
            -060import 
            org.apache.hadoop.hbase.client.Table;
            -061import 
            org.apache.hadoop.hbase.client.TableState;
            -062import 
            org.apache.hadoop.hbase.constraint.ConstraintException;
            -063import 
            org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint;
            -064import 
            org.apache.hadoop.hbase.exceptions.DeserializationException;
            -065import 
            org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
            -066import 
            org.apache.hadoop.hbase.master.MasterServices;
            -067import 
            org.apache.hadoop.hbase.master.ServerListener;
            -068import 
            org.apache.hadoop.hbase.master.TableStateManager;
            -069import 
            org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode;
            -070import 
            org.apache.hadoop.hbase.net.Address;
            -071import 
            org.apache.hadoop.hbase.procedure2.Procedure;
            -072import 
            org.apache.hadoop.hbase.protobuf.ProtobufMagic;
            -073import 
            org.apache.hadoop.hbase.protobuf.ProtobufUtil;
            -074import 
            org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos;
            -075import 
            org.apache.hadoop.hbase.protobuf.generated.RSGroupProtos;
            -076import 
            org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy;
            -077import 
            org.apache.hadoop.hbase.security.access.AccessControlLists;
            -078import 
            org.apache.hadoop.hbase.util.Bytes;
            -079import 
            org.apache.hadoop.hbase.zookeeper.ZKUtil;
            -080import 
            org.apache.hadoop.hbase.zookeeper.ZKWatcher;
            -081import 
            org.apache.hadoop.hbase.zookeeper.ZNodePaths;
            -082import 
            org.apache.yetus.audience.InterfaceAudience;
            -083import 
            org.apache.zookeeper.KeeperException;
            -084import org.slf4j.Logger;
            -085import org.slf4j.LoggerFactory;
            -086
            -087import 
            org.apache.hbase.thirdparty.com.google.common.collect.Lists;
            -088import 
            org.apache.hbase.thirdparty.com.google.common.collect.Maps;
            -089import 
            org.apache.hbase.thirdparty.com.google.common.collect.Sets;
            -090import 
            org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
            -091import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
            -092
            -093/**
            -094 * This is an implementation of {@link 
            RSGroupInfoManager} which makes
            -095 * use of an HBase table as the 
            persistence store for the group information.
            -096 * It also makes use of zookeeper to 
            store group information needed
            -097 * for bootstrapping during offline 
            mode.
            -098 *
            -099 * 

            Concurrency

            -100 * RSGroup state is kept locally in Maps. There is a rsgroup name to cached -101 * RSGroupInfo Map at {@link #rsGroupMap} and a Map of tables to the name of the -102 * rsgroup they belong too (in {@link #tableMap}). These Maps are persisted to the -103 * hbase:rsgroup table (and cached in zk) on each modification. -104 * -105 *

            Mutations on state are synchronized but reads can continue without having -106 * to wait on an instance monitor, mutations do wholesale replace of the Maps on -107 * update -- Copy-On-Write; the local Maps of state are read-only, just-in-case -108 * (see flushConfig). -109 * -110 *

            Reads must not block else there is a danger we'll deadlock. -111 * -112 *

            Clients of this class, the {@link RSGroupAdminEndpoint} for example, want to query and -113 * then act on the results of the query modifying cache in zookeeper without another thread -114 * making intermediate modifications. These clients synchronize on the 'this' instance so -115 * no other has access concurrently. Reads must be able to continue concurrently. -116 */ -117@InterfaceAudience.Private -118final class RSGroupInfoManagerImpl implements RSGroupInfoManager { -119 private static final Logger LOG = LoggerFactory


            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
            --
            diff --git 
            a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
             
            b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
            index bf1e334..f55c3a6 100644
            --- 
            a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
            +++ 
            b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Exemplar for hbase-client archetype – Reactor 
            Dependency Convergence
             
            @@ -912,7 +912,7 @@
             https://www.apache.org/";>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
            --
            diff --git 
            a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
             
            b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
            index bdccdc9..692a5be 100644
            --- 
            a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
            +++ 
            b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Exemplar for hbase-client archetype – 
            Dependency Information
             
            @@ -147,7 +147,7 @@
             https://www.apache.org/";>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
            --
            diff --git 
            a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
             
            b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
            index 482ab8a..04acfbe 100644
            --- 
            a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
            +++ 
            b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Exemplar for hbase-client archetype – Project 
            Dependency Management
             
            @@ -810,7 +810,7 @@
             https://www.apache.org/";>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
            --
            diff --git 
            a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html 
            b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
            index 9bf6e8e..9d5eee1 100644
            --- a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
            +++ b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Exemplar for hbase-client archetype – 
            About
             
            @@ -119,7 +119,7 @@
             https://www.apache.org/";>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
            --
            diff --git 
            a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
             
            b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
            index 06a6c6f..3b2cf47 100644
            --- 
            a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
            +++ 
            b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Exemplar for hbase-client archetype – CI 
            Management
             
            @@ -126,7 +126,7 @@
              

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
            index 0b621b6..69a8636 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
            @@ -82,7 +82,7 @@
             074  }
             075
             076  @Override
            -077  public WAL getWAL(final byte[] 
            identifier, byte[] namespace) throws IOException {
            +077  public WAL getWAL(RegionInfo region) 
            throws IOException {
             078return disabled;
             079  }
             080
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html
            index 0b621b6..69a8636 100644
            --- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html
            +++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html
            @@ -82,7 +82,7 @@
             074  }
             075
             076  @Override
            -077  public WAL getWAL(final byte[] 
            identifier, byte[] namespace) throws IOException {
            +077  public WAL getWAL(RegionInfo region) 
            throws IOException {
             078return disabled;
             079  }
             080
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
            index 4fec5eb..63c9ca7 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
            @@ -35,254 +35,262 @@
             027import java.util.List;
             028import 
            java.util.concurrent.ConcurrentHashMap;
             029import 
            java.util.concurrent.ConcurrentMap;
            -030
            +030import java.util.concurrent.locks.Lock;
             031import 
            org.apache.hadoop.conf.Configuration;
            -032import 
            org.apache.yetus.audience.InterfaceAudience;
            -033import org.slf4j.Logger;
            -034import org.slf4j.LoggerFactory;
            -035// imports for classes still in 
            regionserver.wal
            -036import 
            org.apache.hadoop.hbase.regionserver.wal.WALActionsListener;
            -037import 
            org.apache.hadoop.hbase.util.Bytes;
            -038import 
            org.apache.hadoop.hbase.util.IdLock;
            -039
            -040/**
            -041 * A WAL Provider that returns a WAL per 
            group of regions.
            -042 *
            -043 * This provider follows the decorator 
            pattern and mainly holds the logic for WAL grouping.
            -044 * WAL creation/roll/close is delegated 
            to {@link #DELEGATE_PROVIDER}
            -045 *
            -046 * Region grouping is handled via {@link 
            RegionGroupingStrategy} and can be configured via the
            -047 * property 
            "hbase.wal.regiongrouping.strategy". Current strategy choices are
            -048 * 
              -049 *
            • defaultStrategy : Whatever strategy this version of HBase picks. currently -050 * "bounded".
            • -051 *
            • identity : each region belongs to its own group.
            • -052 *
            • bounded : bounded number of groups and region evenly assigned to each group.
            • -053 *
            -054 * Optionally, a FQCN to a custom implementation may be given. -055 */ -056@InterfaceAudience.Private -057public class RegionGroupingProvider implements WALProvider { -058 private static final Logger LOG = LoggerFactory.getLogger(RegionGroupingProvider.class); -059 -060 /** -061 * Map identifiers to a group number. -062 */ -063 public static interface RegionGroupingStrategy { -064String GROUP_NAME_DELIMITER = "."; -065 -066/** -067 * Given an identifier and a namespace, pick a group. -068 */ -069String group(final byte[] identifier, byte[] namespace); -070void init(Configuration config, String providerId); -071 } -072 -073 /** -074 * Maps between configuration names for strategies and implementation classes. -075 */ -076 static enum Strategies { -077 defaultStrategy(BoundedGroupingStrategy.class), -078 identity(IdentityGroupingStrategy.class), -079 bounded(BoundedGroupingStrategy.class), -080 namespace(NamespaceGroupingStrategy.class); -081 -082final Class clazz; -083Strategies(Class clazz) { -084 this.clazz = clazz; -085} -086 } -087 -088 /** -089 * instant

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
            index 67417b5..900808a 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
            @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private static class HBaseAdmin.AbortProcedureFuture
            +private static class HBaseAdmin.AbortProcedureFuture
             extends HBaseAdmin.ProcedureFutureBoolean>
             
             
            @@ -235,7 +235,7 @@ extends 
             
             isAbortInProgress
            -private boolean isAbortInProgress
            +private boolean isAbortInProgress
             
             
             
            @@ -252,7 +252,7 @@ extends 
             
             AbortProcedureFuture
            -public AbortProcedureFuture(HBaseAdmin admin,
            +public AbortProcedureFuture(HBaseAdmin admin,
             http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true";
             title="class or interface in java.lang">Long procId,
             http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true";
             title="class or interface in 
            java.lang">Boolean abortProcResponse)
             
            @@ -271,7 +271,7 @@ extends 
             
             get
            -public http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true";
             title="class or interface in java.lang">Boolean get(long timeout,
            +public http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true";
             title="class or interface in java.lang">Boolean get(long timeout,
            http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
             title="class or interface in java.util.concurrent">TimeUnit unit)
             throws http://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true";
             title="class or interface in java.lang">InterruptedException,
            http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutionException.html?is-external=true";
             title="class or interface in java.util.concurrent">ExecutionException,
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
            index 4e91404..9d834f8 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
            @@ -132,7 +132,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private static class HBaseAdmin.AddColumnFamilyFuture
            +private static class HBaseAdmin.AddColumnFamilyFuture
             extends HBaseAdmin.ModifyTableFuture
             
             
            @@ -246,7 +246,7 @@ extends 
             
             AddColumnFamilyFuture
            -public AddColumnFamilyFuture(HBaseAdmin admin,
            +public AddColumnFamilyFuture(HBaseAdmin admin,
              TableName tableName,
              
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse response)
             
            @@ -265,7 +265,7 @@ extends 
             
             getOperationType
            -public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String getOperationType()
            +public http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String getOperationType()
             
             Overrides:
             getOperationType in
             class HBaseAdmin.ModifyTableFuture
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
            index 485fa7a..40023f1 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
            @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private static class HBaseAdmin.CreateTableFuture
            +private static class HBaseAdmin.CreateTableFuture
             extends HBaseAdmin.TableFutureVoid>
             
             
            @@ -268,7 +268,7 @@ extends 
             
             desc
            -private final

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html 
            b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
            index 19aa0b9..aedcf26 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
            @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
             
             
             @InterfaceAudience.Private
            -class RawAsyncHBaseAdmin
            +class RawAsyncHBaseAdmin
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object
             implements AsyncAdmin
             The implementation of AsyncAdmin.
            @@ -704,12 +704,12 @@ implements 
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFuture
            -getClusterStatus() 
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFuture
            +getClusterMetrics() 
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFuture
            -getClusterStatus(http://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true";
             title="class or interface in java.util">EnumSet options) 
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFuture
            +getClusterMetrics(http://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true";
             title="class or interface in java.util">EnumSet options) 
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFuture
            @@ -788,27 +788,27 @@ implements 
            -private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            -getRegionLoad(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest request,
            - ServerName serverName) 
            +(package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFuture
            +getRegionLocation(byte[] regionNameOrEncodedRegionName)
            +Get the region location for the passed region name.
            +
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            -getRegionLoads(ServerName serverName)
            -Get a list of RegionLoad of all regions hosted on a 
            region seerver.
            -
            +private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            +getRegionMetrics(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest request,
            +ServerName serverName) 
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            -getRegionLoads(ServerName serverName,
            -  TableName tableName)
            -Get a list of RegionLoad of all regions hosted on a 
            region seerver for a table.
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in java.util.concurrent">CompletableFutureList>
            +getRegionMetrics(ServerName serverName)
            +Get a list of RegionMetrics of 
            all regions hosted on a region seerver.
             
             
             
            -(package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Completab

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
            index 5b3b750..a1f3f7e 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
            @@ -97,3307 +97,3304 @@
             089import 
            org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer;
             090import 
            org.apache.hbase.thirdparty.io.netty.util.Timeout;
             091import 
            org.apache.hbase.thirdparty.io.netty.util.TimerTask;
            -092import 
            org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
            -093import 
            org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
            -094import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
            -095import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
            -096import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse;
            -097import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
            -098import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse;
            -099import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
            -100import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse;
            -101import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest;
            -102import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse;
            -103import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
            -104import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
            -105import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest;
            -106import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse;
            -107import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest;
            -108import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse;
            -109import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest;
            -110import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse;
            -111import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest;
            -112import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse;
            -113import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
            -114import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
            -115import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
            -116import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
            -117import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
            -118import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
            -119import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
            -120import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
            -121import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
            -122import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
            -123import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
            -124import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest;
            -125import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse;
            -126import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
            -127import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
            -128import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest;
            -129import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse;
            -130import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest;
            -131import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse;
            -132import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
            -133import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
            -134impo

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
            index 1fd770d..5c0a3d9 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
            @@ -1961,6 +1961,6 @@ public http://docs.oracle.com/javase/8/docs/api/java/lang/String.h
             
             
             
            -Copyright © 2007–2017 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
            +Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html 
            b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html
            index 1ce7c03..782e4e3 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html
            @@ -1553,6 +1553,6 @@ public Copyright © 2007–2017 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
            +Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html 
            b/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
            index 037cb4a..dd4afb1 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
            @@ -620,6 +620,6 @@ extends Copyright © 2007–2017 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
            +Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html 
            b/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
            index f65cdb2..66f3198 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
            @@ -409,6 +409,6 @@ not permitted.)
             
             
             
            -Copyright © 2007–2017 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
            +Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableState.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableState.html 
            b/devapidocs/org/apache/hadoop/hbase/client/TableState.html
            index 310a067..0ea4ab9 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/TableState.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/TableState.html
            @@ -556,6 +556,6 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             
            -Copyright © 2007–2017 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
            +Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html 
            b/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html
            index 76cec51..6e021bc 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html
            @@ -356,6 +356,6 @@ extends 
             
             
            -Copyright © 2007–2017 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
            +Copyright © 2007–2018 https://www.apache.org/";>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/Vers

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
            index 91eec45..d1cd185 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
            @@ -88,400 +88,396 @@
             080  private final static String 
            CONF_COMPACT_ONCE = "hbase.compactiontool.compact.once";
             081  private final static String 
            CONF_COMPACT_MAJOR = "hbase.compactiontool.compact.major";
             082  private final static String 
            CONF_DELETE_COMPACTED = "hbase.compactiontool.delete";
            -083  private final static String 
            CONF_COMPLETE_COMPACTION = "hbase.hstore.compaction.complete";
            -084
            -085  /**
            -086   * Class responsible to execute the 
            Compaction on the specified path.
            -087   * The path can be a table, region or 
            family directory.
            -088   */
            -089  private static class CompactionWorker 
            {
            -090private final boolean 
            keepCompactedFiles;
            -091private final boolean 
            deleteCompacted;
            -092private final Configuration conf;
            -093private final FileSystem fs;
            -094private final Path tmpDir;
            -095
            -096public CompactionWorker(final 
            FileSystem fs, final Configuration conf) {
            -097  this.conf = conf;
            -098  this.keepCompactedFiles = 
            !conf.getBoolean(CONF_COMPLETE_COMPACTION, true);
            -099  this.deleteCompacted = 
            conf.getBoolean(CONF_DELETE_COMPACTED, false);
            -100  this.tmpDir = new 
            Path(conf.get(CONF_TMP_DIR));
            -101  this.fs = fs;
            -102}
            -103
            -104/**
            -105 * Execute the compaction on the 
            specified path.
            -106 *
            -107 * @param path Directory path on 
            which to run compaction.
            -108 * @param compactOnce Execute just a 
            single step of compaction.
            -109 * @param major Request major 
            compaction.
            -110 */
            -111public void compact(final Path path, 
            final boolean compactOnce, final boolean major) throws IOException {
            -112  if (isFamilyDir(fs, path)) {
            -113Path regionDir = 
            path.getParent();
            -114Path tableDir = 
            regionDir.getParent();
            -115TableDescriptor htd = 
            FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir);
            -116RegionInfo hri = 
            HRegionFileSystem.loadRegionInfoFileContent(fs, regionDir);
            -117compactStoreFiles(tableDir, htd, 
            hri,
            -118path.getName(), compactOnce, 
            major);
            -119  } else if (isRegionDir(fs, path)) 
            {
            -120Path tableDir = 
            path.getParent();
            -121TableDescriptor htd = 
            FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir);
            -122compactRegion(tableDir, htd, 
            path, compactOnce, major);
            -123  } else if (isTableDir(fs, path)) 
            {
            -124compactTable(path, compactOnce, 
            major);
            -125  } else {
            -126throw new IOException(
            -127  "Specified path is not a table, 
            region or family directory. path=" + path);
            -128  }
            -129}
            -130
            -131private void compactTable(final Path 
            tableDir, final boolean compactOnce, final boolean major)
            -132throws IOException {
            -133  TableDescriptor htd = 
            FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir);
            -134  for (Path regionDir: 
            FSUtils.getRegionDirs(fs, tableDir)) {
            -135compactRegion(tableDir, htd, 
            regionDir, compactOnce, major);
            -136  }
            -137}
            -138
            -139private void compactRegion(final Path 
            tableDir, final TableDescriptor htd,
            -140final Path regionDir, final 
            boolean compactOnce, final boolean major)
            -141throws IOException {
            -142  RegionInfo hri = 
            HRegionFileSystem.loadRegionInfoFileContent(fs, regionDir);
            -143  for (Path familyDir: 
            FSUtils.getFamilyDirs(fs, regionDir)) {
            -144compactStoreFiles(tableDir, htd, 
            hri, familyDir.getName(), compactOnce, major);
            -145  }
            -146}
            -147
            -148/**
            -149 * Execute the actual compaction 
            job.
            -150 * If the compact once flag is not 
            specified, execute the compaction until
            -151 * no more compactions are needed. 
            Uses the Configuration settings provided.
            -152 */
            -153private void compactStoreFiles(final 
            Path tableDir, final TableDescriptor htd,
            -154final RegionInfo hri, final 
            String familyName, final boolean compactOnce,
            -155final boolean major) throws 
            IOException {
            -156  HStore store = getStore(conf, fs, 
            tableDir, htd, hri, familyName, tmpDir);
            -157  LOG.info("Compact table=" + 
            htd.getTableName() +
            -158" region=" + 
            hri.getRegionNameAsString() +
            -159" family=" + familyName);
            -160  if (major) {
            -161store.triggerMajorCompaction();
            -162  }
            -163  do {
            -164Optional 
            compaction =
            -165
            store.requestCompa

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html 
            b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
            index edd369a..6fa4fbd 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
            @@ -175,32 +175,32 @@ implements 
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse
            -abortProcedure(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +abortProcedure(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
               
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse
            -addColumn(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +addColumn(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
              
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse
            -addReplicationPeer(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +addReplicationPeer(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
               
            org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse
            -assignRegion(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +assignRegion(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
             
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse
            -balance(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +balance(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
            
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse
            -clearDeadServers(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +clearDeadServers(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
             
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest request) 
             
             
            @@ -209,347 +209,347 @@ implements 
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse
            -createNamespace(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +createNamespace(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
            
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse
            -createTable(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +createTable(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
            
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse
            -decommissionRegionServers(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +decommissionRegionServers(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
              
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse
            -deleteColumn(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +deleteColumn(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
             
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse
            -deleteNamespace(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
            +deleteNamespace(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
            
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest request) 
             
             
             org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse
            -deleteSnapshot(org.a

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html
             
            b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html
            new file mode 100644
            index 000..662952e
            --- /dev/null
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html
            @@ -0,0 +1,1117 @@
            +http://www.w3.org/TR/html4/loose.dtd";>
            +
            +
            +
            +
            +
            +PrivateCellUtil.TagRewriteByteBufferExtendedCell (Apache HBase 
            3.0.0-SNAPSHOT API)
            +
            +
            +
            +
            +
            +var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10};
            +var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
            +var altColor = "altColor";
            +var rowColor = "rowColor";
            +var tableTab = "tableTab";
            +var activeTableTab = "activeTableTab";
            +
            +
            +JavaScript is disabled on your browser.
            +
            +
            +
            +
            +
            +Skip navigation links
            +
            +
            +
            +
            +Overview
            +Package
            +Class
            +Use
            +Tree
            +Deprecated
            +Index
            +Help
            +
            +
            +
            +
            +Prev Class
            +Next Class
            +
            +
            +Frames
            +No Frames
            +
            +
            +All Classes
            +
            +
            +
            +
            +
            +
            +
            +Summary: 
            +Nested | 
            +Field | 
            +Constr | 
            +Method
            +
            +
            +Detail: 
            +Field | 
            +Constr | 
            +Method
            +
            +
            +
            +
            +
            +
            +
            +
            +org.apache.hadoop.hbase
            +Class PrivateCellUtil.TagRewriteByteBufferExtendedCell
            +
            +
            +
            +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">java.lang.Object
            +
            +
            +org.apache.hadoop.hbase.ByteBufferExtendedCell
            +
            +
            +org.apache.hadoop.hbase.PrivateCellUtil.TagRewriteByteBufferExtendedCell
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +All Implemented Interfaces:
            +http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true";
             title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell
            +
            +
            +Direct Known Subclasses:
            +PrivateCellUtil.ValueAndTagRewriteByteBufferExtendedCell
            +
            +
            +Enclosing class:
            +PrivateCellUtil
            +
            +
            +
            +static class PrivateCellUtil.TagRewriteByteBufferExtendedCell
            +extends ByteBufferExtendedCell
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +Nested Class Summary
            +
            +
            +
            +
            +Nested classes/interfaces inherited from 
            interface org.apache.hadoop.hbase.Cell
            +Cell.Type
            +
            +
            +
            +
            +
            +
            +
            +
            +Field Summary
            +
            +Fields 
            +
            +Modifier and Type
            +Field and Description
            +
            +
            +protected ByteBufferExtendedCell
            +cell 
            +
            +
            +private static int
            +HEAP_SIZE_OVERHEAD 
            +
            +
            +protected byte[]
            +tags 
            +
            +
            +
            +
            +
            +
            +Fields inherited from interface org.apache.hadoop.hbase.ExtendedCell
            +CELL_NOT_BASED_ON_CHUNK
            +
            +
            +
            +
            +
            +Fields inherited from interface org.apache.hadoop.hbase.RawCell
            +MAX_TAGS_LENGTH
            +
            +
            +
            +
            +
            +
            +
            +
            +Constructor Summary
            +
            +Constructors 
            +
            +Constructor and Description
            +
            +
            +TagRewriteByteBufferExtendedCell(ByteBufferExtendedCell cell,
            +byte[] tags) 
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +Method Summary
            +
            +All Methods Instance Methods Concrete Methods 
            +
            +Modifier and Type
            +Method and Description
            +
            +
            +ExtendedCell
            +deepClone()
            +Does a deep copy of the contents to a new memory area and 
            returns it as a new cell.
            +
            +
            +
            +byte[]
            +getFamilyArray()
            +Contiguous bytes composed of legal HDFS filename characters 
            which may start at any index in the
            + containing array.
            +
            +
            +
            +http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true";
             title="class or interface in java.nio">ByteBuffer
            +getFamilyByteBuffer() 
            +
            +
            +byte
            +getFamilyLength() 
            +
            +
            +int
            +getFamilyOffset() 
            +
            +
            +int
            +getFamilyPosition() 
            +
            +
            +byte[]
            +getQualifierArray()
            +Contiguous raw bytes that may start at any index in the 
            containing array.
            +
            +
            +
            +http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true";
             title="class or interface in java.nio">ByteBuffer
            +getQualifierByteBuffer() 
            +
            +
            +int
            +getQualifierLength() 
            +
            +
            +int
            +getQualifierOffset() 
            +
            +
            +int
            +getQualifierPosition() 
            +
            +
            +byte[]
            +getRowArray()
            +Contiguous raw bytes that may start at any index in the 
            containing array.
            +
            +
            +
            +http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true";
             title="class or interface in java.nio">ByteBuffer
            +getRowByteBuffer() 
            +
            +
            +short
            +getRowLength() 
            +
            +
            +int
            +get

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
            index 3d37d41..5a14666 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
            @@ -132,7 +132,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private static final class HttpProxyExample.RequestHandler
            +private static final class HttpProxyExample.RequestHandler
             extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHandler
             
             
            @@ -302,7 +302,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             conn
            -private final AsyncConnection conn
            +private final AsyncConnection conn
             
             
             
            @@ -311,7 +311,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             channelGroup
            -private 
            final org.apache.hadoop.hbase.shaded.io.netty.channel.group.ChannelGroup 
            channelGroup
            +private 
            final org.apache.hadoop.hbase.shaded.io.netty.channel.group.ChannelGroup 
            channelGroup
             
             
             
            @@ -328,7 +328,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             RequestHandler
            -public RequestHandler(AsyncConnection conn,
            +public RequestHandler(AsyncConnection conn,
               
            org.apache.hadoop.hbase.shaded.io.netty.channel.group.ChannelGroup channelGroup)
             
             
            @@ -346,7 +346,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             channelActive
            -public void channelActive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx)
            +public void channelActive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx)
             
             Specified by:
             channelActive in 
            interface org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelInboundHandler
            @@ -361,7 +361,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             channelInactive
            -public void channelInactive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx)
            +public void channelInactive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx)
             
             Specified by:
             channelInactive in 
            interface org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelInboundHandler
            @@ -376,7 +376,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             write
            -private void write(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx,
            +private void write(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx,
            
            org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.HttpResponseStatus status,
            http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true";
             title="class or interface in java.util">OptionalString> content)
             
            @@ -387,7 +387,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             parse
            -private HttpProxyExample.Params parse(org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequest req)
            +private HttpProxyExample.Params parse(org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequest req)
             
             
             
            @@ -396,7 +396,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             get
            -private void get(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx,
            +private void get(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx,
              
            org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequest req)
             
             
            @@ -406,7 +406,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             put
            -private void put(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx,
            +private void put(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx,
              
            org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequest req)
             
             
            @@ -416,7 +416,7 @@ extends 
            org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
             
             
             channelRead0
            -protected void channelRead0(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx,
            +protected void channelRead0(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext ctx,
             
            org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequest

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html 
            b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
            index 0c57dc7..061085c 100644
            --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
            +++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
            @@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab";
             
             
             @InterfaceAudience.LimitedPrivate(value="Tools")
            -public class HMaster
            +public class HMaster
             extends HRegionServer
             implements MasterServices
             HMaster is the "master server" for HBase. An HBase cluster 
            has one active
            @@ -1434,7 +1434,7 @@ implements 
             
             LOG
            -private static org.slf4j.Logger LOG
            +private static org.slf4j.Logger LOG
             
             
             
            @@ -1443,7 +1443,7 @@ implements 
             
             MASTER
            -public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String MASTER
            +public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String MASTER
             
             See Also:
             Constant
             Field Values
            @@ -1456,7 +1456,7 @@ implements 
             
             activeMasterManager
            -private final ActiveMasterManager activeMasterManager
            +private final ActiveMasterManager activeMasterManager
             
             
             
            @@ -1465,7 +1465,7 @@ implements 
             
             regionServerTracker
            -RegionServerTracker regionServerTracker
            +RegionServerTracker regionServerTracker
             
             
             
            @@ -1474,7 +1474,7 @@ implements 
             
             drainingServerTracker
            -private DrainingServerTracker drainingServerTracker
            +private DrainingServerTracker drainingServerTracker
             
             
             
            @@ -1483,7 +1483,7 @@ implements 
             
             loadBalancerTracker
            -LoadBalancerTracker loadBalancerTracker
            +LoadBalancerTracker loadBalancerTracker
             
             
             
            @@ -1492,7 +1492,7 @@ implements 
             
             splitOrMergeTracker
            -private SplitOrMergeTracker splitOrMergeTracker
            +private SplitOrMergeTracker splitOrMergeTracker
             
             
             
            @@ -1501,7 +1501,7 @@ implements 
             
             regionNormalizerTracker
            -private RegionNormalizerTracker 
            regionNormalizerTracker
            +private RegionNormalizerTracker 
            regionNormalizerTracker
             
             
             
            @@ -1510,7 +1510,7 @@ implements 
             
             maintenanceModeTracker
            -private MasterMaintenanceModeTracker maintenanceModeTracker
            +private MasterMaintenanceModeTracker maintenanceModeTracker
             
             
             
            @@ -1519,7 +1519,7 @@ implements 
             
             clusterSchemaService
            -private ClusterSchemaService clusterSchemaService
            +private ClusterSchemaService clusterSchemaService
             
             
             
            @@ -1528,7 +1528,7 @@ implements 
             
             HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
            -public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
            +public static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
             
             See Also:
             Constant
             Field Values
            @@ -1541,7 +1541,7 @@ implements 
             
             DEFAULT_HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
            -public static final int DEFAULT_HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
            +public static final int DEFAULT_HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
             
             See Also:
             Constant
             Field Values
            @@ -1554,7 +1554,7 @@ implements 
             
             metricsMaster
            -final MetricsMaster metricsMaster
            +final MetricsMaster metricsMaster
             
             
             
            @@ -1563,7 +1563,7 @@ implements 
             
             fileSystemManager
            -private MasterFileSystem fileSystemManager
            +private MasterFileSystem fileSystemManager
             
             
             
            @@ -1572,7 +1572,7 @@ implements 
             
             walManager
            -private MasterWalManager walManager
            +private MasterWalManager walManager
             
             
             
            @@ -1581,7 +1581,7 @@ implements 
             
             serverManager
            -private volatile ServerManager serverManager
            +private volatile ServerManager serverManager
             
             
             
            @@ -1590,7 +1590,7 @@ implements 
             
             assignmentManager
            -private AssignmentManager assignmentManager
            +private AssignmentManager assignmentManager
             
             
             
            @@ -1599,7 +1599,7 @@ implements 
             
             replicationManager
            -private ReplicationManager replicationManager
            +private ReplicationManager replicationManager
             
             
             
            @@ -1608,7 +1608,7 @@ implements 
             
             rsFatals
            -MemoryBoundedLogMessageBuffer rsFatals
            +MemoryBoundedLogMessageBuffer rsFatals
             
             
             
            @@ -1617,7 +1617,7 @@ implements 
             
             activeMaster
            -private volatile boolean activeMaster
            +private volatile boolean activeMaster
             
             
             
            @@ -1626,7 +1626,7 @@ implements 
             
             initialized
            -private final ProcedureEvent initialized
            +private final ProcedureEvent initialized
             
             
             
            @@ -1635,7 +1635,7 @@ implements 
             
             serviceStarted
            -volatile boolean serviceStarted
            +volatile boolean serviceStarted
             
             
             
            @@ -1644,7 +1644,7 @@ implements 
             
             serverCrashProcessingEnabled
            -private final ProcedureEvent serverCrashProcessingEnabled
            +private final ProcedureEvent serverCrashProcessingEnabled
             
             
             
            @@ -1653,7 +1653,7 @@ implements 
             
             maxBlancingTime
            -p

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html 
            b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
            index c4d3f6e..bec1bbc 100644
            --- a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
            +++ b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
            @@ -191,8 +191,8 @@
             
             java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
             title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
             title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
             title="class or interface in java.io">Serializable)
             
            -org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection
             org.apache.hadoop.hbase.security.AuthMethod
            +org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection
             org.apache.hadoop.hbase.security.SaslStatus
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html 
            b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
            index cd08056..570c4fa 100644
            --- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
            +++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
            @@ -198,9 +198,9 @@
             
             java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
             title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
             title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
             title="class or interface in java.io">Serializable)
             
            +org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType
             org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType
             org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage
            -org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html 
            b/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
            index 43c904b..58a529e 100644
            --- a/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
            +++ b/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
            @@ -320,7 +320,7 @@ extends Admin java 
            class.
              Replication to secondary region replicas works over standard inter-cluster 
            replication.·
             
             See Also:
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html 
            b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
            index bc6e3ef..9aba27a 100644
            --- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
            +++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
            @@ -536,14 +536,14 @@
             
             java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
             title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
             title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
             title="class or interface in java.io">Serializable)
             
            -org.apache.hadoop.hbase.util.ChecksumType
            -org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
            -org.apache.hadoop.hbase.util.PrettyPrinter.Unit
             org.apache.hadoop.hbase.util.Order
            -org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE
            -org.apache.hadoop.hbase.util.PoolMap.PoolType
            +org.apache.hadoop.hbase.util.ChecksumType
             org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer
             (implements org.apache.hadoop.hbase.util.Bytes.Comparer)
            +org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE
             org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer
             (implements org.apache.hadoop.hbase.util.Bytes.Comparer)
            +org.apache.hadoop.hbase.util.PrettyPrinter.Unit
            +org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
            +org.apache.hadoop.hbase.util.PoolMap.PoolType
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
            --
            diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html 
            b/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
            index d7f2c1a..9b468f8 100644
            --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
            +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
            @@ -36,2607 +36,2606 @@
             028import java.util.TreeMap;
             029import java.util.concurrent.Future;
             030import java.util.regex.Pattern;
            -031
            -032import 
            org.apache.hadoop.conf.Configuration;
            -033import 
            org.apache.hadoop.hbase.Abortable;
            -034import 
            org.apache.hadoop.hbase.CacheEvictionStats;
            +031import 
            org.apache.hadoop.conf.Configuration;
            +032import 
            org.apache.hadoop.hbase.Abortable;
            +033import 
            org.apache.hadoop.hbase.CacheEvictionStats;
            +034import 
            org.apache.hadoop.hbase.ClusterMetrics.Option;
             035import 
            org.apache.hadoop.hbase.ClusterStatus;
            -036import 
            org.apache.hadoop.hbase.ClusterStatus.Option;
            -037import 
            org.apache.hadoop.hbase.HRegionInfo;
            -038import 
            org.apache.hadoop.hbase.HTableDescriptor;
            -039import 
            org.apache.hadoop.hbase.NamespaceDescriptor;
            -040import 
            org.apache.hadoop.hbase.NamespaceNotFoundException;
            -041import 
            org.apache.hadoop.hbase.RegionLoad;
            -042import 
            org.apache.hadoop.hbase.ServerName;
            -043import 
            org.apache.hadoop.hbase.TableExistsException;
            -044import 
            org.apache.hadoop.hbase.TableName;
            -045import 
            org.apache.hadoop.hbase.TableNotFoundException;
            -046import 
            org.apache.yetus.audience.InterfaceAudience;
            -047import 
            org.apache.hadoop.hbase.client.replication.TableCFs;
            -048import 
            org.apache.hadoop.hbase.client.security.SecurityCapability;
            -049import 
            org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
            -050import 
            org.apache.hadoop.hbase.quotas.QuotaFilter;
            -051import 
            org.apache.hadoop.hbase.quotas.QuotaRetriever;
            -052import 
            org.apache.hadoop.hbase.quotas.QuotaSettings;
            -053import 
            org.apache.hadoop.hbase.regionserver.wal.FailedLogCloseException;
            -054import 
            org.apache.hadoop.hbase.replication.ReplicationException;
            -055import 
            org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
            -056import 
            org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
            -057import 
            org.apache.hadoop.hbase.snapshot.HBaseSnapshotException;
            -058import 
            org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
            -059import 
            org.apache.hadoop.hbase.snapshot.SnapshotCreationException;
            -060import 
            org.apache.hadoop.hbase.snapshot.UnknownSnapshotException;
            -061import 
            org.apache.hadoop.hbase.util.Bytes;
            -062import 
            org.apache.hadoop.hbase.util.Pair;
            -063
            -064/**
            -065 * The administrative API for HBase. 
            Obtain an instance from {@link Connection#getAdmin()} and
            -066 * call {@link #close()} when done.
            -067 * 

            Admin can be used to create, drop, list, enable and disable and otherwise modify tables, -068 * as well as perform other administrative operations. -069 * -070 * @see ConnectionFactory -071 * @see Connection -072 * @see Table -073 * @since 0.99.0 -074 */ -075@InterfaceAudience.Public -076public interface Admin extends Abortable, Closeable { -077 int getOperationTimeout(); -078 -079 @Override -080 void abort(String why, Throwable e); -081 -082 @Override -083 boolean isAborted(); -084 -085 /** -086 * @return Connection used by this object. -087 */ -088 Connection getConnection(); -089 -090 /** -091 * @param tableName Table to check. -092 * @return true if table exists already. -093 * @throws IOException -094 */ -095 boolean tableExists(TableName tableName) throws IOException; -096 -097 /** -098 * List all the userspace tables. -099 * -100 * @return an array of read-only HTableDescriptors -101 * @throws IOException if a remote or network exception occurs -102 * @deprecated since 2.0 version and will be removed in 3.0 version. -103 * Use {@link #listTableDescriptors()}. -104 * @see #listTableDescriptors() -105 */ -106 @Deprecated -107 HTableDescriptor[] listTables() throws IOException; -108 -109 /** -110 * List all the userspace tables. -111 * -112 * @return a list of TableDescriptors -113 * @throws IOException if a remote or network exception occurs -114 */ -115 List listTableDescriptors() throws IOException; -116 -117 /** -118 * List all the userspace tables that match the given pattern. -119 * -120 * @param pattern The compiled regular expression to match against -121 * @return an array of read-only HTableDescriptors -122 * @throws IOException if a remote or network exception occurs -123 * @see #listTables() -124 * @deprecated since 2.0 version and will be removed in 3.0 version. -125 * Use {@link #listTableDescriptors(java.util.regex.Pattern)}. -126 * @see #listTableDescriptors(Pattern) -127 */ -128 @Deprecated -129 HTableDescriptor[] listTables(Pattern pattern) thro


            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html
            --
            diff --git 
            a/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html 
            b/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html
            index 6c97088..f999cbd 100644
            --- a/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html
            +++ b/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html
            @@ -42,679 +42,678 @@
             034
             035import 
            org.apache.commons.cli.CommandLine;
             036import org.apache.commons.cli.Option;
            -037import org.apache.commons.logging.Log;
            -038import 
            org.apache.commons.logging.LogFactory;
            -039import 
            org.apache.hadoop.conf.Configuration;
            -040import org.apache.hadoop.fs.FileStatus;
            -041import org.apache.hadoop.fs.FileSystem;
            -042import org.apache.hadoop.fs.Path;
            -043import 
            org.apache.hadoop.hbase.TableName;
            -044import 
            org.apache.hadoop.hbase.client.RegionInfo;
            -045import 
            org.apache.hadoop.hbase.client.SnapshotDescription;
            -046import 
            org.apache.hadoop.hbase.io.HFileLink;
            -047import 
            org.apache.hadoop.hbase.io.WALLink;
            -048import 
            org.apache.hadoop.hbase.util.AbstractHBaseTool;
            -049import 
            org.apache.hadoop.hbase.util.FSUtils;
            -050import 
            org.apache.hadoop.util.StringUtils;
            -051import 
            org.apache.yetus.audience.InterfaceAudience;
            -052
            -053import 
            org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
            -054import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos;
            -055import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
            -056
            -057/**
            -058 * Tool for dumping snapshot 
            information.
            -059 * 
              -060 *
            1. Table Descriptor -061 *
            2. Snapshot creation time, type, format version, ... -062 *
            3. List of hfiles and wals -063 *
            4. Stats about hfiles and logs sizes, percentage of shared with the source table, ... -064 *
            -065 */ -066@InterfaceAudience.Public -067public final class SnapshotInfo extends AbstractHBaseTool { -068 private static final Log LOG = LogFactory.getLog(SnapshotInfo.class); -069 -070 static final class Options { -071static final Option SNAPSHOT = new Option(null, "snapshot", true, "Snapshot to examine."); -072static final Option REMOTE_DIR = new Option(null, "remote-dir", true, -073"Root directory that contains the snapshots."); -074static final Option LIST_SNAPSHOTS = new Option(null, "list-snapshots", false, -075"List all the available snapshots and exit."); -076static final Option FILES = new Option(null, "files", false, "Files and logs list."); -077static final Option STATS = new Option(null, "stats", false, "Files and logs stats."); -078static final Option SCHEMA = new Option(null, "schema", false, -079"Describe the snapshotted table."); -080static final Option SIZE_IN_BYTES = new Option(null, "size-in-bytes", false, -081"Print the size of the files in bytes."); -082 } -083 -084 /** -085 * Statistics about the snapshot -086 *
              -087 *
            1. How many store files and logs are in the archive -088 *
            2. How many store files and logs are shared with the table -089 *
            3. Total store files and logs size and shared amount -090 *
            -091 */ -092 public static class SnapshotStats { -093/** Information about the file referenced by the snapshot */ -094static class FileInfo { -095 private final boolean corrupted; -096 private final boolean inArchive; -097 private final long size; -098 -099 FileInfo(final boolean inArchive, final long size, final boolean corrupted) { -100this.corrupted = corrupted; -101this.inArchive = inArchive; -102this.size = size; -103 } -104 -105 /** @return true if the file is in the archive */ -106 public boolean inArchive() { -107return this.inArchive; -108 } -109 -110 /** @return true if the file is corrupted */ -111 public boolean isCorrupted() { -112return this.corrupted; -113 } -114 -115 /** @return true if the file is missing */ -116 public boolean isMissing() { -117return this.size < 0; -118 } -119 -120 /** @return the file size */ -121 public long getSize() { -122return this.size; -123 } -124 -125 String getStateToString() { -126if (isCorrupted()) return "CORRUPTED"; -127if (isMissing()) return "NOT FOUND"; -128if (inArchive()) return "archive"; -129return null; -130 } -131} -132 -133private AtomicInteger hfilesArchiveCount = new AtomicInteger(); -134private AtomicInteger hfilesCorrupted = new AtomicInteger(); -135private AtomicInteger hfilesMissing = new AtomicInteger(); -136private AtomicInteger hfilesCount = new AtomicInteger(); -137private AtomicInteger hfilesMobCount = new AtomicInteger(); -138private AtomicInteger logsMissing

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d0f1a9f6/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
            index e354b98..d875570 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
            @@ -129,117 +129,116 @@
             121   */
             122  @Override
             123  public void execute() throws 
            IOException {
            -124try (Admin admin = conn.getAdmin();) 
            {
            -125
            -126  // Begin BACKUP
            -127  beginBackup(backupManager, 
            backupInfo);
            -128  String savedStartCode = null;
            -129  boolean firstBackup = false;
            -130  // do snapshot for full table 
            backup
            -131
            -132  savedStartCode = 
            backupManager.readBackupStartCode();
            -133  firstBackup = savedStartCode == 
            null || Long.parseLong(savedStartCode) == 0L;
            -134  if (firstBackup) {
            -135// This is our first backup. 
            Let's put some marker to system table so that we can hold the logs
            -136// while we do the backup.
            -137
            backupManager.writeBackupStartCode(0L);
            -138  }
            -139  // We roll log here before we do 
            the snapshot. It is possible there is duplicate data
            -140  // in the log that is already in 
            the snapshot. But if we do it after the snapshot, we
            -141  // could have data loss.
            -142  // A better approach is to do the 
            roll log on each RS in the same global procedure as
            -143  // the snapshot.
            -144  LOG.info("Execute roll log 
            procedure for full backup ...");
            -145
            -146  Map props = 
            new HashMap();
            -147  props.put("backupRoot", 
            backupInfo.getBackupRootDir());
            -148  
            admin.execProcedure(LogRollMasterProcedureManager.ROLLLOG_PROCEDURE_SIGNATURE,
            -149
            LogRollMasterProcedureManager.ROLLLOG_PROCEDURE_NAME, props);
            -150
            -151  newTimestamps = 
            backupManager.readRegionServerLastLogRollResult();
            -152  if (firstBackup) {
            -153// Updates registered log files
            -154// We record ALL old WAL files as 
            registered, because
            -155// this is a first full backup in 
            the system and these
            -156// files are not needed for next 
            incremental backup
            -157List logFiles = 
            BackupUtils.getWALFilesOlderThan(conf, newTimestamps);
            -158
            backupManager.recordWALFiles(logFiles);
            -159  }
            -160
            -161  // SNAPSHOT_TABLES:
            -162  
            backupInfo.setPhase(BackupPhase.SNAPSHOT);
            -163  for (TableName tableName : 
            tableList) {
            -164String snapshotName =
            -165"snapshot_" + 
            Long.toString(EnvironmentEdgeManager.currentTime()) + "_"
            -166+ 
            tableName.getNamespaceAsString() + "_" + tableName.getQualifierAsString();
            -167
            -168snapshotTable(admin, tableName, 
            snapshotName);
            -169
            backupInfo.setSnapshotName(tableName, snapshotName);
            -170  }
            -171
            -172  // SNAPSHOT_COPY:
            -173  // do snapshot copy
            -174  LOG.debug("snapshot copy for " + 
            backupId);
            -175  snapshotCopy(backupInfo);
            -176  // Updates incremental backup table 
            set
            -177  
            backupManager.addIncrementalBackupTableSet(backupInfo.getTables());
            -178
            -179  // BACKUP_COMPLETE:
            -180  // set overall backup status: 
            complete. Here we make sure to complete the backup.
            -181  // After this checkpoint, even if 
            entering cancel process, will let the backup finished
            -182  
            backupInfo.setState(BackupState.COMPLETE);
            -183  // The table list in backupInfo is 
            good for both full backup and incremental backup.
            -184  // For incremental backup, it 
            contains the incremental backup table set.
            -185  
            backupManager.writeRegionServerLogTimestamp(backupInfo.getTables(), 
            newTimestamps);
            -186
            -187  HashMap> newTableSetTimestampMap =
            -188  
            backupManager.readLogTimestampMap();
            -189
            -190  Long newStartCode =
            -191  
            BackupUtils.getMinValue(BackupUtils
            -192  
            .getRSLogTimestampMins(newTableSetTimestampMap));
            -193  
            backupManager.writeBackupStartCode(newStartCode);
            -194
            -195  // backup complete
            -196  completeBackup(conn, backupInfo, 
            backupManager, BackupType.FULL, conf);
            -197} catch (Exception e) {
            -198  failBackup(conn, backupInfo, 
            backupManager, e, "Unexpected BackupException : ",
            -199BackupType.FULL, conf);
            -200  throw new IOException(e);
            -201}
            -202
            -203  }
            +124try (Admin admin = conn.getAdmin()) 
            {
            +125  // Begin BACKUP
            +126  beginBackup(backupManager, 
            backupInfo);
            +127  String savedStartCode = null;
            +128  boolean firstBackup = false;
            +129  // do snapshot for full table 
            backup
            +130
            +131  savedStartCode = 
            backupManager.readBackupStartCode();
            +132  firstBackup = savedSta

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dad9a249/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.html 
            b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.html
            index 83099f8..9c5b56f 100644
            --- a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.html
            +++ b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9,"i40":9,"i41":9,"i42":9,"i43":9,"i44":9,"i45":9,"i46":9,"i47":9,"i48":9,"i49":9,"i50":9,"i51":9,"i52":9,"i53":9,"i54":9,"i55":9,"i56":9,"i57":9,"i58":9,"i59":9,"i60":9,"i61":9,"i62":9,"i63":9,"i64":9,"i65":9,"i66":9,"i67":9,"i68":9,"i69":9,"i70":9,"i71":9,"i72":9,"i73":9,"i74":9,"i75":9,"i76":9,"i77":9,"i78":9,"i79":9,"i80":9,"i81":9,"i82":9,"i83":9,"i84":9,"i85":9,"i86":9,"i87":9,"i88":9,"i89":9,"i90":9};
            +var methods = 
            {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9,"i40":9,"i41":9,"i42":9,"i43":9,"i44":9,"i45":9,"i46":9,"i47":9,"i48":9,"i49":9,"i50":9,"i51":9,"i52":9,"i53":9,"i54":9,"i55":9,"i56":9,"i57":9,"i58":9,"i59":9,"i60":9,"i61":9,"i62":9,"i63":9,"i64":9,"i65":9,"i66":9,"i67":9,"i68":9,"i69":9,"i70":9,"i71":9,"i72":9,"i73":9,"i74":9,"i75":9,"i76":9,"i77":9,"i78":9,"i79":9,"i80":9,"i81":9,"i82":9,"i83":9,"i84":9,"i85":9,"i86":9,"i87":9,"i88":9,"i89":9,"i90":9,"i91":9,"i92":9};
             var tabs = {65535:["t0","All Methods"],1:["t1","Static 
            Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
             
             
             @InterfaceAudience.Private
            -public final class PrivateCellUtil
            +public final class PrivateCellUtil
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object
             Utility methods helpful slinging Cell instances. It has more powerful 
            and
              rich set of APIs than those in CellUtil for internal usage.
            @@ -723,7 +723,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             
            -private static http://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true";
             title="class or interface in java.util">Iterator
            +static http://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true";
             title="class or interface in java.util">Iterator
             tagsIterator(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true";
             title="class or interface in java.nio">ByteBuffer tags,
             int offset,
             int length) 
            @@ -735,6 +735,14 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             
            +static Cell.DataType
            +toDataType(byte type) 
            +
            +
            +static KeyValue.Type
            +toTypeByte(Cell.DataType type) 
            +
            +
             static boolean
             updateLatestStamp(Cell cell,
              byte[] ts)
            @@ -742,7 +750,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
              HConstants.LATEST_TIMESTAMP.
             
             
            -
            +
             static boolean
             updateLatestStamp(Cell cell,
              long ts)
            @@ -750,7 +758,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
              HConstants.LATEST_TIMESTAMP.
             
             
            -
            +
             static int
             writeCell(Cell cell,
              http://docs.oracle.com/javase/8/docs/api/java/io/OutputStream.html?is-external=true";
             title="class or interface in java.io">OutputStream out,
            @@ -758,7 +766,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             Writes the cell to the given OutputStream
             
             
            -
            +
             static void
             writeCellToBuffer(Cell cell,
              http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true";
             title="class or interface in java.nio">ByteBuffer buf,
            @@ -766,7 +774,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             Writes a cell to the buffer at the given offset
             
             
            -
            +
             static void
             writeFamily(http://docs.oracle.com/javase/8/docs/api/java/io/OutputStream.html?is-external=true";
             title="class or interface in java.io">OutputStream out,
            Cell cell,
            @@ -774,19 +782,19 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             Writes the family from the given cell to the output 
            stream
             
             
            -
            +
             static void
             writeFlatKey(Cell cell,
             http://docs.oracle.com/javase/8/docs/api/java/io/DataOutput.html?is-external=true

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
            --
            diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html 
            b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
            index 4dedbc2..ffeadbf 100644
            --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
            +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
            @@ -902,350 +902,351 @@
             894return allowPartialResults;
             895  }
             896
            -897  public Scan 
            setLoadColumnFamiliesOnDemand(boolean value) {
            -898return (Scan) 
            super.setLoadColumnFamiliesOnDemand(value);
            -899  }
            -900
            -901  /**
            -902   * Compile the table and column family 
            (i.e. schema) information
            -903   * into a String. Useful for parsing 
            and aggregation by debugging,
            -904   * logging, and administration tools.
            -905   * @return Map
            -906   */
            -907  @Override
            -908  public Map 
            getFingerprint() {
            -909Map map = new 
            HashMap<>();
            -910List families = new 
            ArrayList<>();
            -911if(this.familyMap.isEmpty()) {
            -912  map.put("families", "ALL");
            -913  return map;
            -914} else {
            -915  map.put("families", families);
            -916}
            -917for (Map.Entry> entry :
            -918this.familyMap.entrySet()) {
            -919  
            families.add(Bytes.toStringBinary(entry.getKey()));
            -920}
            -921return map;
            -922  }
            -923
            -924  /**
            -925   * Compile the details beyond the scope 
            of getFingerprint (row, columns,
            -926   * timestamps, etc.) into a Map along 
            with the fingerprinted information.
            -927   * Useful for debugging, logging, and 
            administration tools.
            -928   * @param maxCols a limit on the number 
            of columns output prior to truncation
            -929   * @return Map
            -930   */
            -931  @Override
            -932  public Map 
            toMap(int maxCols) {
            -933// start with the fingerpring map and 
            build on top of it
            -934Map map = 
            getFingerprint();
            -935// map from families to column list 
            replaces fingerprint's list of families
            -936Map> 
            familyColumns = new HashMap<>();
            -937map.put("families", familyColumns);
            -938// add scalar information first
            -939map.put("startRow", 
            Bytes.toStringBinary(this.startRow));
            -940map.put("stopRow", 
            Bytes.toStringBinary(this.stopRow));
            -941map.put("maxVersions", 
            this.maxVersions);
            -942map.put("batch", this.batch);
            -943map.put("caching", this.caching);
            -944map.put("maxResultSize", 
            this.maxResultSize);
            -945map.put("cacheBlocks", 
            this.cacheBlocks);
            -946map.put("loadColumnFamiliesOnDemand", 
            this.loadColumnFamiliesOnDemand);
            -947List timeRange = new 
            ArrayList<>(2);
            -948timeRange.add(this.tr.getMin());
            -949timeRange.add(this.tr.getMax());
            -950map.put("timeRange", timeRange);
            -951int colCount = 0;
            -952// iterate through affected families 
            and list out up to maxCols columns
            -953for (Map.Entry> entry :
            -954  this.familyMap.entrySet()) {
            -955  List columns = new 
            ArrayList<>();
            -956  
            familyColumns.put(Bytes.toStringBinary(entry.getKey()), columns);
            -957  if(entry.getValue() == null) {
            -958colCount++;
            -959--maxCols;
            -960columns.add("ALL");
            -961  } else {
            -962colCount += 
            entry.getValue().size();
            -963if (maxCols <= 0) {
            -964  continue;
            -965}
            -966for (byte [] column : 
            entry.getValue()) {
            -967  if (--maxCols <= 0) {
            -968continue;
            -969  }
            -970  
            columns.add(Bytes.toStringBinary(column));
            -971}
            -972  }
            -973}
            -974map.put("totalColumns", colCount);
            -975if (this.filter != null) {
            -976  map.put("filter", 
            this.filter.toString());
            -977}
            -978// add the id if set
            -979if (getId() != null) {
            -980  map.put("id", getId());
            -981}
            -982return map;
            -983  }
            -984
            -985  /**
            -986   * Enable/disable "raw" mode for this 
            scan.
            -987   * If "raw" is enabled the scan will 
            return all
            -988   * delete marker and deleted rows that 
            have not
            -989   * been collected, yet.
            -990   * This is mostly useful for Scan on 
            column families
            -991   * that have KEEP_DELETED_ROWS 
            enabled.
            -992   * It is an error to specify any column 
            when "raw" is set.
            -993   * @param raw True/False to 
            enable/disable "raw" mode.
            -994   */
            -995  public Scan setRaw(boolean raw) {
            -996setAttribute(RAW_ATTR, 
            Bytes.toBytes(raw));
            -997return this;
            -998  }
            -999
            -1000  /**
            -1001   * @return True if this Scan is in 
            "raw" mode.
            -1002   */
            -1003  public boolean isRaw() {
            -1004byte[] attr = 
            getAttribute(RAW_ATTR);
            -1005return attr == null ? false : 
            Bytes.toBoolean(attr);
            -1006  }
            -1007
            -1008  /**
            -1009   * Set whether this scan is a small 
            scan
            -1010   * 

            -1011 * Small scan should use pread and big scan can use seek + read seek + read is fast but can cause -1012 * two prob


            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
             
            b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
            index 54a1767..ff18b00 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
            @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -public static class ThriftServerRunner.HBaseHandler
            +public static class ThriftServerRunner.HBaseHandler
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
             title="class or interface in java.lang">Object
             implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             The HBaseHandler is a glue object that connects Thrift RPC 
            calls to the
            @@ -721,7 +721,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             conf
            -protected org.apache.hadoop.conf.Configuration conf
            +protected org.apache.hadoop.conf.Configuration conf
             
             
             
            @@ -730,7 +730,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             LOG
            -protected static final org.apache.commons.logging.Log LOG
            +protected static final org.apache.commons.logging.Log LOG
             
             
             
            @@ -739,7 +739,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             nextScannerId
            -protected int nextScannerId
            +protected int nextScannerId
             
             
             
            @@ -748,7 +748,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             scannerMap
            -protected http://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true";
             title="class or interface in java.util">HashMapInteger,ThriftServerRunner.ResultScannerWrapper> 
            scannerMap
            +protected http://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true";
             title="class or interface in java.util">HashMapInteger,ThriftServerRunner.ResultScannerWrapper> 
            scannerMap
             
             
             
            @@ -757,7 +757,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             metrics
            -private ThriftMetrics metrics
            +private ThriftMetrics metrics
             
             
             
            @@ -766,7 +766,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             connectionCache
            -private final ConnectionCache connectionCache
            +private final ConnectionCache connectionCache
             
             
             
            @@ -775,7 +775,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             coalescer
            -IncrementCoalescer coalescer
            +IncrementCoalescer coalescer
             
             
             
            @@ -784,7 +784,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             CLEANUP_INTERVAL
            -static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String CLEANUP_INTERVAL
            +static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String CLEANUP_INTERVAL
             
             See Also:
             Constant
             Field Values
            @@ -797,7 +797,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             MAX_IDLETIME
            -static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String MAX_IDLETIME
            +static final http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String MAX_IDLETIME
             
             See Also:
             Constant
             Field Values
            @@ -818,7 +818,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             HBaseHandler
            -protected HBaseHandler(org.apache.hadoop.conf.Configuration c,
            +protected HBaseHandler(org.apache.hadoop.conf.Configuration c,
            UserProvider userProvider)
             throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in java.io">IOException
             
            @@ -841,7 +841,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             getAllColumns
            -byte[][] getAllColumns(Table table)
            +byte[][] getAllColumns(Table table)
             throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in java.io">IOException
             Returns a list of all the column families for a given 
            Table.
             
            @@ -858,7 +858,7 @@ implements 
            org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
             
             
             getTable
            -public Table getTable(byte[] tableName)
            +public Table getTable(byte[] tableName)
            throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
            index 7c59e27..c904c56 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
            @@ -119,4048 +119,4054 @@
             111import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
             112import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
             113import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearRegionBlockCacheRequest;
            -114import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
            -115import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
            -116import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
            -117import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
            -118import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest;
            -119import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse;
            -120import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest;
            -121import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest;
            -122import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
            -123import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest;
            -124import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse;
            -125import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
            -126import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
            -127import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
            -128import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
            -129import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos;
            -130import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
            -131import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
            -132import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
            -133import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
            -134import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
            -135import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest;
            -136import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
            -137import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
            -138import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest;
            -139import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse;
            -140import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
            -141import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
            -142import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
            -143import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse;
            -144import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest;
            -145import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest;
            -146import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse;
            -147import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
            -148import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
            -149import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
            -150import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
            -151import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest;
            -152import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse;
            -153import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest;
            -154import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest;
            -155import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksReque

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/devapidocs/org/apache/hadoop/hbase/HConstants.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/HConstants.html 
            b/devapidocs/org/apache/hadoop/hbase/HConstants.html
            index 8bf6d22..2340381 100644
            --- a/devapidocs/org/apache/hadoop/hbase/HConstants.html
            +++ b/devapidocs/org/apache/hadoop/hbase/HConstants.html
            @@ -623,279 +623,283 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
            -DEFAULT_TEMPORARY_HDFS_DIRECTORY 
            +DEFAULT_STATUS_MULTICAST_PUBLISHER_BIND_ADDRESS 
             
             
            +static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
            +DEFAULT_TEMPORARY_HDFS_DIRECTORY 
            +
            +
             static int
             DEFAULT_THREAD_WAKE_FREQUENCY
             Default value for thread wake frequency
             
             
            -
            +
             static boolean
             DEFAULT_USE_META_REPLICAS 
             
            -
            +
             static int
             DEFAULT_VERSION_FILE_WRITE_ATTEMPTS
             Parameter name for how often we should try to write a 
            version file, before failing
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             DEFAULT_WAL_STORAGE_POLICY 
             
            -
            +
             static int
             DEFAULT_ZK_SESSION_TIMEOUT
             Default value for ZooKeeper session timeout
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             DEFAULT_ZOOKEEPER_ZNODE_PARENT 
             
            -
            +
             static int
             DEFAULT_ZOOKEPER_CLIENT_PORT
             Default client port that the zookeeper listens on
             
             
            -
            +
             static int
             DEFAULT_ZOOKEPER_MAX_CLIENT_CNXNS
             Default limit on concurrent client-side zookeeper 
            connections
             
             
            -
            +
             static int
             DELIMITER
             delimiter used between portions of a region name
             
             
            -
            +
             static byte[]
             EMPTY_BYTE_ARRAY
             An empty instance.
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true";
             title="class or interface in java.nio">ByteBuffer
             EMPTY_BYTE_BUFFER 
             
            -
            +
             static byte[]
             EMPTY_END_ROW
             Last row in a table.
             
             
            -
            +
             static byte[]
             EMPTY_START_ROW
             Used by scanners, etc when they want to start at the 
            beginning of a region
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             ENABLE_CLIENT_BACKPRESSURE
             Config key for if the server should send backpressure and 
            if the client should listen to
              that backpressure from the server
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             ENABLE_DATA_FILE_UMASK
             Enable file permission modification from standard 
            hbase
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             ENABLE_WAL_COMPRESSION
             Configuration name of WAL Compression
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             ENABLE_WAL_ENCRYPTION
             Configuration key for enabling WAL encryption, a 
            boolean
             
             
            -
            +
             static TableName
             ENSEMBLE_TABLE_NAME
             The name of the ensemble table
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             FILE_SYSTEM_VERSION
             Current version of file system.
             
             
            -
            +
             static int
             FOREVER
             Unlimited time-to-live.
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             HBASE_BALANCER_MAX_BALANCING
             Config for the max balancing time
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             HBASE_BALANCER_PERIOD
             Config for the balancer period
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             HBASE_CANARY_READ_RAW_SCAN_KEY 
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             HBASE_CANARY_WRITE_DATA_TTL_KEY
             Canary config keys
             
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             HBASE_CANARY_WRITE_PERSERVER_REGIONS_LOWERLIMIT_KEY 
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String
             HBASE_CANARY_WRITE_PERSERVER_REGIONS_UPPERLIMIT_KEY 
             
            -
            +
             static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lan

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html 
            b/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
            index ed18c40..eb9af9e 100644
            --- a/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
            +++ b/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":41,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":41,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":42,"i40":42,"i41":41,"i42":42,"i43":42,"i44":42,"i45":41,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42,"i67":42,"i68":42,"i69":42,"i70":42,"i71":42,"i72":42,"i73":42,"i74":42,"i75":42,"i76":42,"i77":42,"i78":42,"i79":42,"i80":42,"i81":42,"i82":42,"i83":42};
            +var methods = 
            {"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":41,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":41,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":42,"i40":41,"i41":42,"i42":42,"i43":42,"i44":41,"i45":42,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42,"i67":42,"i68":42,"i69":42,"i70":42,"i71":42,"i72":42,"i73":42,"i74":42,"i75":42,"i76":42,"i77":42,"i78":42,"i79":42,"i80":42,"i81":42,"i82":42};
             var tabs = {65535:["t0","All Methods"],1:["t1","Static 
            Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
            Methods"],32:["t6","Deprecated Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -792,41 +792,35 @@ implements 
             boolean
            -isCacheDataInL1()
            -Deprecated. 
            - 
            -
            -
            -boolean
             isCacheDataOnWrite()
             Deprecated. 
              
             
            -
            +
             boolean
             isCacheIndexesOnWrite()
             Deprecated. 
              
             
            -
            +
             boolean
             isCompressTags()
             Deprecated. 
              
             
            -
            +
             boolean
             isEvictBlocksOnClose()
             Deprecated. 
              
             
            -
            +
             boolean
             isInMemory()
             Deprecated. 
              
             
            -
            +
             static byte[]
             isLegalFamilyName(byte[] b)
             Deprecated. 
            @@ -834,109 +828,113 @@ implements 
            +
             boolean
             isMobEnabled()
             Deprecated. 
             Gets whether the mob is enabled for the family.
             
             
            -
            +
             boolean
             isNewVersionBehavior()
             Deprecated. 
             By default, HBase only consider timestamp in versions.
             
             
            -
            +
             boolean
             isPrefetchBlocksOnOpen()
             Deprecated. 
              
             
            -
            +
             static HColumnDescriptor
             parseFrom(byte[] bytes)
             Deprecated. 
              
             
            -
            +
             void
             remove(byte[] key)
             Deprecated. 
              
             
            -
            +
             void
             removeConfiguration(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String key)
             Deprecated. 
             Remove a configuration setting represented by the key.
             
             
            -
            +
             HColumnDescriptor
             setBlockCacheEnabled(boolean value)
             Deprecated. 
              
             
            -
            +
             HColumnDescriptor
             setBlocksize(int value)
             Deprecated. 
              
             
            -
            +
             HColumnDescriptor
             setBloomFilterType(BloomType value)
             Deprecated. 
              
             
            -
            +
             HColumnDescriptor
             setCacheBloomsOnWrite(boolean value)
             Deprecated. 
              
             
            -
            +
             HColumnDescriptor
             setCacheDataInL1(boolean value)
            -Deprecated. 
            - 
            +Deprecated. 
            +Since 2.0 and will be 
            removed in 3.0 with out any replacement. Caching data in on
            + heap Cache, when there are both on heap LRU Cache and Bucket 
            Cache will no longer
            + be supported from 2.0.
            +
            +
             
            -
            +
             HColumnDescriptor
             setCacheDataOnWrite(boolean value)
             Deprecated. 
              
             
            -
            +
             HColumnDescriptor
             setCacheIndexesOnWrite(boolean value)
             Deprecated. 
              
             
            -
            +
             HColumnDescriptor
             setCompactionCompressionType(Compression.Algorithm value)
             Deprecated. 
             Compression types supported in hbase.
             
             
            -
            +
             HColumnDescriptor
             setCompressionType(Compression.Algorithm value)
             Deprecated. 
             Compression types supported in hbase.
             
             
            -
            +
             HColumnDescriptor
             setCompressTags(boolean value)
             Deprecated. 
             Set whether the tags should be compressed along with 
            DataBlockEncoding.
             
             
            -
            +
             HColumnDescriptor
             setConfiguration(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String key,
             http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
             title="class or interface in java.lang">String value)
            @@ -944,143 +942,143 @@ implements Setter for storing a configuration setting.
             
             
            -
            +
             HColumnDescriptor
             setDataBlockEncoding(DataBlockEncoding 

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html
            --
            diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html
            index 00db109..9c710d6 100644
            --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html
            +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html
            @@ -36,32 +36,32 @@
             028import java.util.NavigableMap;
             029import java.util.TreeMap;
             030import java.util.UUID;
            -031
            -032import org.apache.hadoop.hbase.Cell;
            -033import 
            org.apache.hadoop.hbase.CellScannable;
            -034import 
            org.apache.hadoop.hbase.CellScanner;
            -035import 
            org.apache.hadoop.hbase.CellUtil;
            -036import 
            org.apache.hadoop.hbase.HConstants;
            -037import 
            org.apache.hadoop.hbase.KeyValue;
            -038import 
            org.apache.hadoop.hbase.PrivateCellUtil;
            -039import org.apache.hadoop.hbase.Tag;
            -040import 
            org.apache.hadoop.hbase.exceptions.DeserializationException;
            -041import 
            org.apache.hadoop.hbase.io.HeapSize;
            -042import 
            org.apache.hadoop.hbase.protobuf.ProtobufUtil;
            -043import 
            org.apache.hadoop.hbase.protobuf.generated.ClientProtos;
            -044import 
            org.apache.hadoop.hbase.security.access.AccessControlConstants;
            -045import 
            org.apache.hadoop.hbase.security.access.AccessControlUtil;
            -046import 
            org.apache.hadoop.hbase.security.access.Permission;
            -047import 
            org.apache.hadoop.hbase.security.visibility.CellVisibility;
            -048import 
            org.apache.hadoop.hbase.security.visibility.VisibilityConstants;
            -049import 
            org.apache.hadoop.hbase.shaded.com.google.common.collect.ArrayListMultimap;
            -050import 
            org.apache.hadoop.hbase.shaded.com.google.common.collect.ListMultimap;
            -051import 
            org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataInput;
            -052import 
            org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataOutput;
            -053import 
            org.apache.hadoop.hbase.shaded.com.google.common.io.ByteStreams;
            -054import 
            org.apache.hadoop.hbase.util.Bytes;
            -055import 
            org.apache.hadoop.hbase.util.ClassSize;
            -056import 
            org.apache.yetus.audience.InterfaceAudience;
            +031import org.apache.hadoop.hbase.Cell;
            +032import 
            org.apache.hadoop.hbase.CellScannable;
            +033import 
            org.apache.hadoop.hbase.CellScanner;
            +034import 
            org.apache.hadoop.hbase.CellUtil;
            +035import 
            org.apache.hadoop.hbase.HConstants;
            +036import 
            org.apache.hadoop.hbase.KeyValue;
            +037import 
            org.apache.hadoop.hbase.PrivateCellUtil;
            +038import org.apache.hadoop.hbase.Tag;
            +039import 
            org.apache.hadoop.hbase.exceptions.DeserializationException;
            +040import 
            org.apache.hadoop.hbase.io.HeapSize;
            +041import 
            org.apache.hadoop.hbase.protobuf.ProtobufUtil;
            +042import 
            org.apache.hadoop.hbase.protobuf.generated.ClientProtos;
            +043import 
            org.apache.hadoop.hbase.security.access.AccessControlConstants;
            +044import 
            org.apache.hadoop.hbase.security.access.AccessControlUtil;
            +045import 
            org.apache.hadoop.hbase.security.access.Permission;
            +046import 
            org.apache.hadoop.hbase.security.visibility.CellVisibility;
            +047import 
            org.apache.hadoop.hbase.security.visibility.VisibilityConstants;
            +048import 
            org.apache.hadoop.hbase.util.Bytes;
            +049import 
            org.apache.hadoop.hbase.util.ClassSize;
            +050import 
            org.apache.yetus.audience.InterfaceAudience;
            +051
            +052import 
            org.apache.hadoop.hbase.shaded.com.google.common.collect.ArrayListMultimap;
            +053import 
            org.apache.hadoop.hbase.shaded.com.google.common.collect.ListMultimap;
            +054import 
            org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataInput;
            +055import 
            org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataOutput;
            +056import 
            org.apache.hadoop.hbase.shaded.com.google.common.io.ByteStreams;
             057
             058@InterfaceAudience.Public
             059public abstract class Mutation extends 
            OperationWithAttributes implements Row, CellScannable,
            @@ -223,360 +223,508 @@
             215if (getTTL() != Long.MAX_VALUE) {
             216  map.put("ttl", getTTL());
             217}
            -218return map;
            -219  }
            -220
            -221  private static Map cellToStringMap(Cell c) {
            -222Map stringMap = 
            new HashMap<>();
            -223stringMap.put("qualifier", 
            Bytes.toStringBinary(c.getQualifierArray(), c.getQualifierOffset(),
            -224
            c.getQualifierLength()));
            -225stringMap.put("timestamp", 
            c.getTimestamp());
            -226stringMap.put("vlen", 
            c.getValueLength());
            -227List tags = 
            PrivateCellUtil.getTags(c);
            -228if (tags != null) {
            -229  List tagsString = new 
            ArrayList<>(tags.size());
            -230  for (Tag t : tags) {
            -231tagsString
            -232.add((t.getType()) + ":" + 
            Bytes.toStringBinary(Tag.cloneValue(t)));
            -233  }
            -234  stringMap.put("tag", tagsString);
            -235}
            -236return stringMap;
            -237  }
            -238
            -239  /**
            -240   * Set the durability for this 
            mutation
            -241   * @param d
            -242   */
            -243  public Mutation 
            setDurability(Durability d) {
            -244this.durability = d;
            -245return this

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
            index 3edfbef..9707b2c 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
            @@ -2459,5936 +2459,5935 @@
             2451  }
             2452
             2453  for (HStore s : storesToFlush) {
            -2454MemStoreSize flushableSize = 
            s.getFlushableSize();
            -2455
            totalSizeOfFlushableStores.incMemStoreSize(flushableSize);
            -2456
            storeFlushCtxs.put(s.getColumnFamilyDescriptor().getName(),
            -2457  
            s.createFlushContext(flushOpSeqId, tracker));
            -2458// for writing stores to WAL
            -2459
            committedFiles.put(s.getColumnFamilyDescriptor().getName(), null);
            -2460
            storeFlushableSize.put(s.getColumnFamilyDescriptor().getName(), 
            flushableSize);
            -2461  }
            -2462
            -2463  // write the snapshot start to 
            WAL
            -2464  if (wal != null && 
            !writestate.readOnly) {
            -2465FlushDescriptor desc = 
            ProtobufUtil.toFlushDescriptor(FlushAction.START_FLUSH,
            -2466getRegionInfo(), 
            flushOpSeqId, committedFiles);
            -2467// No sync. Sync is below where 
            no updates lock and we do FlushAction.COMMIT_FLUSH
            -2468WALUtil.writeFlushMarker(wal, 
            this.getReplicationScope(), getRegionInfo(), desc, false,
            -2469mvcc);
            -2470  }
            -2471
            -2472  // Prepare flush (take a 
            snapshot)
            -2473  for (StoreFlushContext flush : 
            storeFlushCtxs.values()) {
            -2474flush.prepare();
            -2475  }
            -2476} catch (IOException ex) {
            -2477  doAbortFlushToWAL(wal, 
            flushOpSeqId, committedFiles);
            -2478  throw ex;
            -2479} finally {
            -2480  
            this.updatesLock.writeLock().unlock();
            -2481}
            -2482String s = "Finished memstore 
            snapshotting " + this + ", syncing WAL and waiting on mvcc, " +
            -2483"flushsize=" + 
            totalSizeOfFlushableStores;
            -2484status.setStatus(s);
            -2485doSyncOfUnflushedWALChanges(wal, 
            getRegionInfo());
            -2486return new 
            PrepareFlushResult(storeFlushCtxs, committedFiles, storeFlushableSize, 
            startTime,
            -2487flushOpSeqId, flushedSeqId, 
            totalSizeOfFlushableStores);
            -2488  }
            -2489
            -2490  /**
            -2491   * Utility method broken out of 
            internalPrepareFlushCache so that method is smaller.
            -2492   */
            -2493  private void 
            logFatLineOnFlush(Collection storesToFlush, long sequenceId) {
            -2494if (!LOG.isInfoEnabled()) {
            -2495  return;
            -2496}
            -2497// Log a fat line detailing what is 
            being flushed.
            -2498StringBuilder perCfExtras = null;
            -2499if (!isAllFamilies(storesToFlush)) 
            {
            -2500  perCfExtras = new 
            StringBuilder();
            -2501  for (HStore store: storesToFlush) 
            {
            -2502perCfExtras.append("; 
            ").append(store.getColumnFamilyName());
            -2503perCfExtras.append("=")
            -2504
            .append(StringUtils.byteDesc(store.getFlushableSize().getDataSize()));
            -2505  }
            -2506}
            -2507LOG.info("Flushing " + + 
            storesToFlush.size() + "/" + stores.size() +
            -2508" column families, memstore=" + 
            StringUtils.byteDesc(this.memstoreDataSize.get()) +
            -2509((perCfExtras != null && 
            perCfExtras.length() > 0)? perCfExtras.toString(): "") +
            -2510((wal != null) ? "" : "; WAL is 
            null, using passed sequenceid=" + sequenceId));
            -2511  }
            -2512
            -2513  private void doAbortFlushToWAL(final 
            WAL wal, final long flushOpSeqId,
            -2514  final Map> committedFiles) {
            -2515if (wal == null) return;
            -2516try {
            -2517  FlushDescriptor desc = 
            ProtobufUtil.toFlushDescriptor(FlushAction.ABORT_FLUSH,
            -2518  getRegionInfo(), flushOpSeqId, 
            committedFiles);
            -2519  WALUtil.writeFlushMarker(wal, 
            this.getReplicationScope(), getRegionInfo(), desc, false,
            -2520  mvcc);
            -2521} catch (Throwable t) {
            -2522  LOG.warn("Received unexpected 
            exception trying to write ABORT_FLUSH marker to WAL:" +
            -2523  
            StringUtils.stringifyException(t));
            -2524  // ignore this since we will be 
            aborting the RS with DSE.
            -2525}
            -2526// we have called 
            wal.startCacheFlush(), now we have to abort it
            -2527
            wal.abortCacheFlush(this.getRegionInfo().getEncodedNameAsBytes());
            -2528  }
            -2529
            -2530  /**
            -2531   * Sync unflushed WAL changes. See 
            HBASE-8208 for details
            -2532   */
            -2533  private static void 
            doSyncOfUnflushedWALChanges(final WAL wal, final RegionInfo hri)
            -2534  throws IOException {
            -2535if (wal == null) {
            -2536  return;
            -2537}
            -2538try {
            -2539  wal.sync(); // ensure that flush 
            marker is sync'ed
            -2540} catch (IOException ioe) {
            -2541  
            wal.abortCacheFlush(hri.getEncodedNameAsBytes());
            -2542 

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html 
            b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
            index e571d5c..3803d9a 100644
            --- a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
            +++ b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
            @@ -206,12 +206,12 @@
             
             java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
             title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
             title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
             title="class or interface in java.io">Serializable)
             
            -org.apache.hadoop.hbase.quotas.ThrottlingException.Type
             org.apache.hadoop.hbase.quotas.OperationQuota.OperationType
            -org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
            -org.apache.hadoop.hbase.quotas.QuotaType
             org.apache.hadoop.hbase.quotas.QuotaScope
             org.apache.hadoop.hbase.quotas.ThrottleType
            +org.apache.hadoop.hbase.quotas.QuotaType
            +org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
            +org.apache.hadoop.hbase.quotas.ThrottlingException.Type
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
             
            b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
            index b091edd..ca8621b 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
            @@ -400,7 +400,7 @@ extends 
             
             triggerFlushInPrimaryRegion
            -void triggerFlushInPrimaryRegion(HRegion region)
            +void triggerFlushInPrimaryRegion(HRegion region)
               throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
             title="class or interface in java.io">IOException,
              http://docs.oracle.com/javase/8/docs/api/java/lang/RuntimeException.html?is-external=true";
             title="class or interface in java.lang">RuntimeException
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
            index c99a865..85b1f93 100644
            --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
            +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
            @@ -699,19 +699,19 @@
             
             java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
             title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
             title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
             title="class or interface in java.io">Serializable)
             
            -org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
            -org.apache.hadoop.hbase.regionserver.BloomType
            -org.apache.hadoop.hbase.regionserver.Region.Operation
             org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result
            -org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
            -org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
            -org.apache.hadoop.hbase.regionserver.FlushType
            -org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
             org.apache.hadoop.hbase.regionserver.ScanType
            +org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
             org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type
            +org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
            +org.apache.hadoop.hbase.regionserver.BloomType
            +org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
            +org.apache.hadoop.hbase.regionserver.Region.Operation
            +org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
            +org.apache.hadoop.hbase.regionserver.FlushType
             org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
             org.apache.hadoop.hbase.regionserver.MemStoreCompactionStrategy.Action
            -org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
            +org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/

            [27/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html 
            b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html
            index 91bfe40..3ddf027 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html
            @@ -279,7 +279,7 @@ implements 
              http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">ListCompletableFuture>
             batch(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List actions)
            -Method that does a batch call on Deletes, Gets, Puts, 
            Increments and Appends.
            +Method that does a batch call on Deletes, Gets, Puts, 
            Increments, Appends and RowMutations.
             
             
             
            @@ -1165,15 +1165,15 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">ListCompletableFuture> batch(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
             title="class or interface in java.util">List actions)
             Description copied from 
            interface: AsyncTable
            -Method that does a batch call on Deletes, Gets, Puts, 
            Increments and Appends. The ordering of
            - execution of the actions is not defined. Meaning if you do a Put and a Get in 
            the same
            - AsyncTable.batch(java.util.List) call, you will not 
            necessarily be guaranteed that the Get returns what the Put
            - had put.
            +Method that does a batch call on Deletes, Gets, Puts, 
            Increments, Appends and RowMutations. The
            + ordering of execution of the actions is not defined. Meaning if you do a Put 
            and a Get in the
            + same AsyncTable.batch(java.util.List) call, you will not 
            necessarily be guaranteed that the Get returns what the
            + Put had put.
             
             Specified by:
             batch in
             interface AsyncTable
             Parameters:
            -actions - list of Get, Put, Delete, Increment, Append 
            objects
            +actions - list of Get, Put, Delete, Increment, Append, and 
            RowMutations objects
             Returns:
             A list of http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
             title="class or interface in 
            java.util.concurrent">CompletableFutures that represent the 
            result for each action.
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html 
            b/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html
            index 495b3bf..66c3237 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":17,"i1":6,"i2":6,"i3":17,"i4":17,"i5":17,"i6":17,"i7":17,"i8":17,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":17,"i18":17,"i19":6,"i20":17,"i21":6,"i22":17,"i23":17,"i24":17,"i25":6,"i26":6,"i27":6,"i28":6,"i29":17,"i30":17,"i31":17,"i32":17,"i33":17,"i34":17,"i35":17,"i36":17,"i37":17,"i38":17,"i39":17};
            +var methods = 
            {"i0":17,"i1":6,"i2":6,"i3":17,"i4":17,"i5":17,"i6":17,"i7":17,"i8":17,"i9":17,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":17,"i19":17,"i20":6,"i21":17,"i22":6,"i23":17,"i24":17,"i25":17,"i26":6,"i27":6,"i28":6,"i29":6,"i30":17,"i31":17,"i32":17,"i33":17,"i34":17,"i35":17,"i36":17,"i37":17,"i38":17,"i39":17,"i40":17};
             var tabs = {65535:["t0","All Methods"],1:["t1","Static 
            Methods"],2:["t2","Instance Methods"],4:["t3","Abstract 
            Methods"],16:["t5","Default Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -212,6 +212,12 @@ public interface containsRow(byte[] row) 
             
             
            +static RegionInfo
            +createMobRegionInfo(TableName tableName)
            +Creates a RegionInfo object for MOB data.
            +
            +
            +
             static byte[]
             createRegionName(TableName tableName,
             byte[] startKey,
            @@ -220,7 +226,7 @@ public interface Make a region name of passed parameters.
             
             
            -
            +
             static byte[]
             createRegionName(TableName tableName,
             byte[] startKey,
            @@ -230,7 +236,7 @@ public interface Make a region name of passed parameters.
             
             
            -
            +
             static byte[]
             createRegionName(TableName tableName,
             

              1   2   3   4   >