[26/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.

2019-01-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
index 81fc69e..6c58b9a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
@@ -47,7 +47,7 @@
 039import 
org.apache.hadoop.hbase.client.Admin;
 040import 
org.apache.hadoop.hbase.client.Connection;
 041import 
org.apache.hadoop.hbase.procedure2.store.wal.WALProcedureStore;
-042import 
org.apache.hadoop.hbase.util.FSUtils;
+042import 
org.apache.hadoop.hbase.util.CommonFSUtils;
 043import 
org.apache.hadoop.hbase.wal.AbstractFSWALProvider;
 044import 
org.apache.yetus.audience.InterfaceAudience;
 045import org.slf4j.Logger;
@@ -251,156 +251,157 @@
 243  throws IOException {
 244LOG.debug("In 
getLogFilesForNewBackup()\n" + "olderTimestamps: " + olderTimestamps
 245+ "\n newestTimestamps: " + 
newestTimestamps);
-246Path rootdir = 
FSUtils.getRootDir(conf);
-247Path logDir = new Path(rootdir, 
HConstants.HREGION_LOGDIR_NAME);
-248Path oldLogDir = new Path(rootdir, 
HConstants.HREGION_OLDLOGDIR_NAME);
-249FileSystem fs = 
rootdir.getFileSystem(conf);
-250NewestLogFilter pathFilter = new 
NewestLogFilter();
-251
-252ListString resultLogFiles = 
new ArrayList();
-253ListString newestLogs = new 
ArrayList();
-254
-255/*
-256 * The old region servers and 
timestamps info we kept in backup system table may be out of sync
-257 * if new region server is added or 
existing one lost. We'll deal with it here when processing
-258 * the logs. If data in backup system 
table has more hosts, just ignore it. If the .logs
-259 * directory includes more hosts, the 
additional hosts will not have old timestamps to compare
-260 * with. We'll just use all the logs 
in that directory. We always write up-to-date region server
-261 * and timestamp info to backup 
system table at the end of successful backup.
-262 */
-263FileStatus[] rss;
-264Path p;
-265String host;
-266Long oldTimeStamp;
-267String currentLogFile;
-268long currentLogTS;
-269
-270// Get the files in .logs.
-271rss = fs.listStatus(logDir);
-272for (FileStatus rs : rss) {
-273  p = rs.getPath();
-274  host = 
BackupUtils.parseHostNameFromLogFile(p);
-275  if (host == null) {
-276continue;
-277  }
-278  FileStatus[] logs;
-279  oldTimeStamp = 
olderTimestamps.get(host);
-280  // It is possible that there is no 
old timestamp in backup system table for this host if
-281  // this region server is newly 
added after our last backup.
-282  if (oldTimeStamp == null) {
-283logs = fs.listStatus(p);
-284  } else {
-285
pathFilter.setLastBackupTS(oldTimeStamp);
-286logs = fs.listStatus(p, 
pathFilter);
-287  }
-288  for (FileStatus log : logs) {
-289LOG.debug("currentLogFile: " + 
log.getPath().toString());
-290if 
(AbstractFSWALProvider.isMetaFile(log.getPath())) {
-291  if (LOG.isDebugEnabled()) {
-292LOG.debug("Skip hbase:meta 
log file: " + log.getPath().getName());
-293  }
-294  continue;
-295}
-296currentLogFile = 
log.getPath().toString();
-297
resultLogFiles.add(currentLogFile);
-298currentLogTS = 
BackupUtils.getCreationTime(log.getPath());
-299
-300// If newestTimestamps.get(host) 
is null, means that
-301// either RS (host) has been 
restarted recently with different port number
-302// or RS is down (was 
decommisioned). In any case, we treat this
-303// log file as eligible for 
inclusion into incremental backup log list
-304Long ts = 
newestTimestamps.get(host);
-305if (ts ==  null) {
-306  LOG.warn("ORPHAN log found: " + 
log + " host=" + host);
-307  LOG.debug("Known hosts (from 
newestTimestamps):");
-308  for (String s: 
newestTimestamps.keySet()) {
-309LOG.debug(s);
-310  }
-311}
-312if (ts == null || currentLogTS 
 ts) {
-313  
newestLogs.add(currentLogFile);
-314}
-315  }
-316}
-317
-318// Include the .oldlogs files too.
-319FileStatus[] oldlogs = 
fs.listStatus(oldLogDir);
-320for (FileStatus oldlog : oldlogs) {
-321  p = oldlog.getPath();
-322  currentLogFile = p.toString();
-323  if 
(AbstractFSWALProvider.isMetaFile(p)) {
-324if (LOG.isDebugEnabled()) {
-325  LOG.debug("Skip .meta log file: 
" + currentLogFile);
-326}
-327continue;
-328  }
-329  

[26/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.

2019-01-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
index d1a64c3..26212bf 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
@@ -288,7 +288,7 @@ implements MasterObserver
-postAbortProcedure,
 postAddReplicationPeer,
 postAddRSGroup,
 postAssign,
 postBalance, postBalanceRSGroup,
 postBalanceSwitch,
 postClearDeadServers,
 postCloneSnapshot,
 postCompletedCreateTableAction,
 postCompletedDeleteTableAction,
 postCompletedDisableTableAction,
 postCompletedEnableTableAction,
 postCompletedMergeRegionsAction,
 postCompletedModifyTableAction,
 postCompletedModifyTableAction,
 postCompletedSplitRegionAction,
 postCompletedTruncateTableAction,
 postCreateNamespace,
 postDecommissionRegionServers,
 postDeleteNamespace, postDeleteSnapshot,
 postDeleteTable,
 postDisableReplicationPeer,
 postDisableTable,
 postEnableReplicationPeer,
 postEnableTable,
 postGetClusterMetrics,
 postGetLocks,
 postGetNamespaceDescriptor,
 postGetProcedures,
 postGetReplicationPeerConfig,
 postGetRSGroupInfo,
 postGetRSGroupInfoOfServer,
 postGetRSGroupInfoOfTable, postGetTableDescriptors,
 postGetTableNames,
 postListDecommissionedRegionServers,
 postListNamespaceDescriptors,
 postListReplicationPeers,
 postListRSGroups,
 postListSnapshot,
 postLockHeartbeat,
 postMergeRegions,
 postMergeRegionsCommitAction,
 postModifyNamespace,
 postModifyNamespace,
 postModifyTable,
 postModifyTable,
 postMove,
 postMoveServers,
 postMoveServersAndTables,
 postMoveTables,
 postRecommissionRegionServer,
 postRegionOffline,
 postRemoveReplicationPeer,
 postRemoveRSGroup, postRemoveServers,
 postRequestLock,
 postRestoreSnapshot,
 postRollBackMergeRegionsAction, postRollBackSplitRegionAction,
 postSetNamespaceQuota,
 postSetSplitOrMergeEnabled,
 pos
 tSetTableQuota, postSetUserQuota,
 postSetUserQuota,
 postSetUserQuota,
 postSnapshot, postStartMaster,
 postTableFlush,
 postTransitReplicationPeerSyncReplicationState,
 postTruncateTable, postUnassign,
 postUpdateReplicationPeerConfig,
 preAbortProcedure,
 preAddReplicationPeer,
 preAddRSGroup,
 preAssign,
 preBalance,
 preBalanceRSGroup,
 preBalanceSwitch,
 preClearDeadServers,
 preCloneSnapshot,
 preCreateNamespace,
 preCreateTableAction,
 preCreateTableRegionsInfos,
 preDecommissionRegionServers,
 preDeleteNamespace,
 preDeleteSnapshot,
 preDeleteTable,
 preDeleteTableAction,
 preDisableReplicationPeer,
 preDisableTableAction,
 preEnableReplicationPeer,
 preEnableTable, preEnableTableAction,
 preGetClusterMetrics,
 preGetLocks,
 preGetNamespaceDescriptor,
 preGetProcedures,
  preGetReplicationPeerConfig,
 preGetRSGroupInfo,
 preGetRSGroupInfoOfServer,
 preGetRSGroupInfoOfTable,
 preGetTableDescriptors,
 preGetTableNames,
 preListDecommissionedRegionServers,
 preListNamespaceDescriptors,
 preListReplicationPeers,
 preListRSGroups,
 preListSnapshot,
 preLockHeartbeat,
 preMasterInitialization,
 preMergeRegions,
 preMergeRegionsAction,
 preMergeRegionsCommitAction,
 preModifyNamespace,
 preModifyNamespace,
 preModifyTable,
 preModifyTable,
 preModifyTableAction,
 preModifyTableAction,
 preMove,
 preMoveServers,
 preMoveServersAndTables,
 preMove
 Tables, preRecommissionRegionServer,
 preRegionOffline,
 preRemoveReplicationPeer,
 preRemoveRSGroup,
 preRemoveServers, preRequestLock,
 preRestoreSnapshot,
 preSetNamespaceQuota,
 preSetSplitOrMergeEnabled,
 preSetTableQuota,
 preSetUserQuota,
 preSetUserQuota,
 preSetUserQuota,
 preShutdown,
 preSnapshot,
 preSplitRegion,
 preSplitRegionAction, preSplitRegionAfterMETAAction,
 preSplitRegionBeforeMETAAction,
 preStopMaster,
 preTableFlush,
 preTransitReplicationPeerSyncReplicationState,
 preTruncateTable,
 preTruncateTableAction,
 preUnassign,
 preUpdateReplicationPeerConfig
+postAbortProcedure,
 postAddReplicationPeer,
 postAddRSGroup,
 postAssign,
 postBalance, postBalanceRSGroup,
 postBalanceSwitch,
 postClearDeadServers,
 postCloneSnapshot,
 postCompletedCreateTableAction,
 postCompletedDeleteTableAction,
 postCompletedDisableTableAction,
 postCompletedEnableTableAction,
 postCompletedMergeRegionsAction,
 postCompletedModifyTableAction,
 postCompletedModifyTableAction,
 postCompletedSplitRegionAction,
 postCompletedTruncateTableAction,
 postCreateNamespace,
 postDecommissionRegionServers,
 postDeleteNamespace, postDeleteSnapshot,
 postDeleteTable,
 postDisableReplicationPeer,
 postDisableTable,
 postEnableReplicationPeer,
 postEnableTable,
 postGetClusterMetrics,
 postGetLocks,
 postGetNamespaceDescriptor,
 postGetProcedures,
 

[26/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.

2019-01-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncConnectionConfiguration.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncConnectionConfiguration.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncConnectionConfiguration.html
index 779b02a..e91f4fe 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncConnectionConfiguration.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncConnectionConfiguration.html
@@ -47,149 +47,161 @@
 039import static 
org.apache.hadoop.hbase.HConstants.HBASE_RPC_WRITE_TIMEOUT_KEY;
 040import static 
org.apache.hadoop.hbase.client.AsyncProcess.DEFAULT_START_LOG_ERRORS_AFTER_COUNT;
 041import static 
org.apache.hadoop.hbase.client.AsyncProcess.START_LOG_ERRORS_AFTER_COUNT_KEY;
-042import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.WRITE_BUFFER_PERIODIC_FLUSH_TIMEOUT_MS;
-043import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.WRITE_BUFFER_PERIODIC_FLUSH_TIMEOUT_MS_DEFAULT;
-044import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.WRITE_BUFFER_SIZE_DEFAULT;
-045import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.WRITE_BUFFER_SIZE_KEY;
-046
-047import java.util.concurrent.TimeUnit;
-048import 
org.apache.hadoop.conf.Configuration;
-049import 
org.apache.hadoop.hbase.HBaseConfiguration;
-050import 
org.apache.yetus.audience.InterfaceAudience;
-051
-052/**
-053 * Timeout configs.
-054 */
-055@InterfaceAudience.Private
-056class AsyncConnectionConfiguration {
-057
-058  private final long 
metaOperationTimeoutNs;
+042import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.PRIMARY_CALL_TIMEOUT_MICROSECOND;
+043import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.PRIMARY_CALL_TIMEOUT_MICROSECOND_DEFAULT;
+044import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.WRITE_BUFFER_PERIODIC_FLUSH_TIMEOUT_MS;
+045import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.WRITE_BUFFER_PERIODIC_FLUSH_TIMEOUT_MS_DEFAULT;
+046import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.WRITE_BUFFER_SIZE_DEFAULT;
+047import static 
org.apache.hadoop.hbase.client.ConnectionConfiguration.WRITE_BUFFER_SIZE_KEY;
+048
+049import java.util.concurrent.TimeUnit;
+050import 
org.apache.hadoop.conf.Configuration;
+051import 
org.apache.hadoop.hbase.HBaseConfiguration;
+052import 
org.apache.yetus.audience.InterfaceAudience;
+053
+054/**
+055 * Timeout configs.
+056 */
+057@InterfaceAudience.Private
+058class AsyncConnectionConfiguration {
 059
-060  // timeout for a whole operation such 
as get, put or delete. Notice that scan will not be effected
-061  // by this value, see scanTimeoutNs.
-062  private final long 
operationTimeoutNs;
-063
-064  // timeout for each rpc request. Can be 
overridden by a more specific config, such as
-065  // readRpcTimeout or writeRpcTimeout.
-066  private final long rpcTimeoutNs;
-067
-068  // timeout for each read rpc request
-069  private final long readRpcTimeoutNs;
-070
-071  // timeout for each write rpc request
-072  private final long writeRpcTimeoutNs;
-073
-074  private final long pauseNs;
+060  private final long 
metaOperationTimeoutNs;
+061
+062  // timeout for a whole operation such 
as get, put or delete. Notice that scan will not be effected
+063  // by this value, see scanTimeoutNs.
+064  private final long 
operationTimeoutNs;
+065
+066  // timeout for each rpc request. Can be 
overridden by a more specific config, such as
+067  // readRpcTimeout or writeRpcTimeout.
+068  private final long rpcTimeoutNs;
+069
+070  // timeout for each read rpc request
+071  private final long readRpcTimeoutNs;
+072
+073  // timeout for each write rpc request
+074  private final long writeRpcTimeoutNs;
 075
-076  private final int maxRetries;
+076  private final long pauseNs;
 077
-078  /** How many retries are allowed before 
we start to log */
-079  private final int startLogErrorsCnt;
-080
-081  // As now we have heartbeat support for 
scan, ideally a scan will never timeout unless the RS is
-082  // crash. The RS will always return 
something before the rpc timeout or scan timeout to tell the
-083  // client that it is still alive. The 
scan timeout is used as operation timeout for every
-084  // operations in a scan, such as 
openScanner or next.
-085  private final long scanTimeoutNs;
-086
-087  private final int scannerCaching;
+078  private final int maxRetries;
+079
+080  /** How many retries are allowed before 
we start to log */
+081  private final int startLogErrorsCnt;
+082
+083  // As now we have heartbeat support for 
scan, ideally a scan will never timeout unless the RS is
+084  // crash. The RS will always return 
something before the rpc timeout or scan timeout to tell the
+085  // client that it is still alive. The 

[26/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.

2019-01-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperationWithResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperationWithResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperationWithResult.html
index c4e8c8b..aa58108 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperationWithResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperationWithResult.html
@@ -82,602 +82,613 @@
 074  public static final String 
USER_COPROCESSORS_ENABLED_CONF_KEY =
 075"hbase.coprocessor.user.enabled";
 076  public static final boolean 
DEFAULT_USER_COPROCESSORS_ENABLED = true;
-077
-078  private static final Logger LOG = 
LoggerFactory.getLogger(CoprocessorHost.class);
-079  protected Abortable abortable;
-080  /** Ordered set of loaded coprocessors 
with lock */
-081  protected final SortedListE 
coprocEnvironments =
-082  new SortedList(new 
EnvironmentPriorityComparator());
-083  protected Configuration conf;
-084  // unique file prefix to use for local 
copies of jars when classloading
-085  protected String pathPrefix;
-086  protected AtomicInteger loadSequence = 
new AtomicInteger();
-087
-088  public CoprocessorHost(Abortable 
abortable) {
-089this.abortable = abortable;
-090this.pathPrefix = 
UUID.randomUUID().toString();
-091  }
-092
-093  /**
-094   * Not to be confused with the 
per-object _coprocessors_ (above),
-095   * coprocessorNames is static and 
stores the set of all coprocessors ever
-096   * loaded by any thread in this JVM. It 
is strictly additive: coprocessors are
-097   * added to coprocessorNames, by 
checkAndLoadInstance() but are never removed, since
-098   * the intention is to preserve a 
history of all loaded coprocessors for
-099   * diagnosis in case of server crash 
(HBASE-4014).
-100   */
-101  private static SetString 
coprocessorNames =
-102  Collections.synchronizedSet(new 
HashSetString());
-103
-104  public static SetString 
getLoadedCoprocessors() {
-105synchronized (coprocessorNames) {
-106  return new 
HashSet(coprocessorNames);
-107}
-108  }
-109
-110  /**
-111   * Used to create a parameter to the 
HServerLoad constructor so that
-112   * HServerLoad can provide information 
about the coprocessors loaded by this
-113   * regionserver.
-114   * (HBASE-4070: Improve region server 
metrics to report loaded coprocessors
-115   * to master).
-116   */
-117  public SetString 
getCoprocessors() {
-118SetString returnValue = new 
TreeSet();
-119for (E e: coprocEnvironments) {
-120  
returnValue.add(e.getInstance().getClass().getSimpleName());
-121}
-122return returnValue;
-123  }
-124
-125  /**
-126   * Load system coprocessors once only. 
Read the class names from configuration.
-127   * Called by constructor.
-128   */
-129  protected void 
loadSystemCoprocessors(Configuration conf, String confKey) {
-130boolean coprocessorsEnabled = 
conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
-131  DEFAULT_COPROCESSORS_ENABLED);
-132if (!coprocessorsEnabled) {
-133  return;
-134}
-135
-136Class? implClass;
-137
-138// load default coprocessors from 
configure file
-139String[] defaultCPClasses = 
conf.getStrings(confKey);
-140if (defaultCPClasses == null || 
defaultCPClasses.length == 0)
-141  return;
-142
-143int priority = 
Coprocessor.PRIORITY_SYSTEM;
-144for (String className : 
defaultCPClasses) {
-145  className = className.trim();
-146  if (findCoprocessor(className) != 
null) {
-147// If already loaded will just 
continue
-148LOG.warn("Attempted duplicate 
loading of " + className + "; skipped");
-149continue;
-150  }
-151  ClassLoader cl = 
this.getClass().getClassLoader();
-152  
Thread.currentThread().setContextClassLoader(cl);
-153  try {
-154implClass = 
cl.loadClass(className);
-155// Add coprocessors as we go to 
guard against case where a coprocessor is specified twice
-156// in the configuration
-157E env = 
checkAndLoadInstance(implClass, priority, conf);
-158if (env != null) {
-159  
this.coprocEnvironments.add(env);
-160  LOG.info("System coprocessor {} 
loaded, priority={}.", className, priority);
-161  ++priority;
-162}
-163  } catch (Throwable t) {
-164// We always abort if system 
coprocessors cannot be loaded
-165abortServer(className, t);
-166  }
-167}
-168  }
-169
-170  /**
-171   * Load a coprocessor implementation 
into the host
-172   * @param path path to implementation 
jar
-173   * @param className the main class 
name
-174   * @param priority chaining priority

[26/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.

2019-01-01 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/PreemptiveFastFailInterceptor.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/PreemptiveFastFailInterceptor.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/PreemptiveFastFailInterceptor.html
index c459751..1447281 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/PreemptiveFastFailInterceptor.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/PreemptiveFastFailInterceptor.html
@@ -120,6 +120,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/Put.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Put.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/Put.html
index d2bac64..e03fc85 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Put.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Put.html
@@ -1469,6 +1469,6 @@ service.
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/Query.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Query.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/Query.html
index b11cbf3..1347704 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Query.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Query.html
@@ -283,6 +283,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/QuotaStatusCalls.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/QuotaStatusCalls.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/QuotaStatusCalls.html
index 5e74280..c041959 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/QuotaStatusCalls.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/QuotaStatusCalls.html
@@ -120,6 +120,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
index b86c21f..0d59b19 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
@@ -120,6 +120,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AdminRpcCall.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AdminRpcCall.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AdminRpcCall.html
index deae215..66cbac7 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AdminRpcCall.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.AdminRpcCall.html
@@ -166,6 +166,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 


[26/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.

2018-12-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html 
b/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
index cd76f45..98b72a3 100644
--- a/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
+++ b/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":9,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109
 
":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10};
+var methods = 
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":9,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109
 
":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.LimitedPrivate(value="Configuration")
-public class AccessController
+public class AccessController
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements MasterCoprocessor, RegionCoprocessor, RegionServerCoprocessor, 
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.Interface,
 MasterObserver, 
RegionObserver, 
RegionServerObserver, EndpointObserver, 
BulkLoadObserver
 Provides basic authorization checks for data access and 
administrative
@@ -376,72 +376,78 @@ implements 
+private Cell
+createNewCellWithTags(Mutationmutation,
+ CelloldCell,
+ CellnewCell)
+
+
 private User
 getActiveUser(ObserverContext?ctx)
 Returns the active user to which authorization checks 
should be applied.
 
 
-
+
 AuthManager
 getAuthManager()
 
-
+
 https://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalBulkLoadObserver
 getBulkLoadObserver()
 
-
+
 https://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalEndpointObserver
 getEndpointObserver()
 
-
+
 https://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalMasterObserver
 getMasterObserver()
 
-
+
 Region
 getRegion()
 

[26/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.

2018-12-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
index cd14215..07384a0 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":9,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":9,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":9,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109
 
":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":9,"i131":10,"i132":9,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":9,"i152":10,"i153":9,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":9,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":9,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":9,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109
 
":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":9,"i132":10,"i133":9,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":9,"i153":10,"i154":9,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.LimitedPrivate(value="Tools")
-public class HRegionServer
+public class HRegionServer
 extends HasThread
 implements RegionServerServices, LastSequenceId, 
ConfigurationObserver
 HRegionServer makes a set of HRegions available to clients. 
It checks in with
@@ -205,8 +205,8 @@ implements abortRequested
 
 
-protected CacheConfig
-cacheConfig
+private BlockCache
+blockCache
 
 
 protected MemStoreFlusher
@@ -359,8 +359,8 @@ implements metricsTable
 
 
-(package private) MobCacheConfig
-mobCacheConfig
+private MobFileCache
+mobFileCache
 
 
 protected https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,HRegionServer.MovedRegionInfo
@@ -867,8 +867,10 @@ implements 

[26/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.

2018-12-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
index ea05301..26a93dd 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
@@ -269,3590 +269,3574 @@
 261   */
 262  protected ClusterConnection 
clusterConnection;
 263
-264  /*
-265   * Long-living meta table locator, 
which is created when the server is started and stopped
-266   * when server shuts down. References 
to this locator shall be used to perform according
-267   * operations in EventHandlers. Primary 
reason for this decision is to make it mockable
-268   * for tests.
-269   */
-270  protected MetaTableLocator 
metaTableLocator;
-271
-272  /**
-273   * Go here to get table descriptors.
-274   */
-275  protected TableDescriptors 
tableDescriptors;
-276
-277  // Replication services. If no 
replication, this handler will be null.
-278  protected ReplicationSourceService 
replicationSourceHandler;
-279  protected ReplicationSinkService 
replicationSinkHandler;
-280
-281  // Compactions
-282  public CompactSplit 
compactSplitThread;
-283
-284  /**
-285   * Map of regions currently being 
served by this region server. Key is the
-286   * encoded region name.  All access 
should be synchronized.
-287   */
-288  protected final MapString, 
HRegion onlineRegions = new ConcurrentHashMap();
-289
-290  /**
-291   * Map of encoded region names to the 
DataNode locations they should be hosted on
-292   * We store the value as 
InetSocketAddress since this is used only in HDFS
-293   * API (create() that takes favored 
nodes as hints for placing file blocks).
-294   * We could have used ServerName here 
as the value class, but we'd need to
-295   * convert it to InetSocketAddress at 
some point before the HDFS API call, and
-296   * it seems a bit weird to store 
ServerName since ServerName refers to RegionServers
-297   * and here we really mean DataNode 
locations.
-298   */
-299  protected final MapString, 
InetSocketAddress[] regionFavoredNodesMap =
-300  new ConcurrentHashMap();
-301
-302  // Leases
-303  protected Leases leases;
+264  /**
+265   * Go here to get table descriptors.
+266   */
+267  protected TableDescriptors 
tableDescriptors;
+268
+269  // Replication services. If no 
replication, this handler will be null.
+270  protected ReplicationSourceService 
replicationSourceHandler;
+271  protected ReplicationSinkService 
replicationSinkHandler;
+272
+273  // Compactions
+274  public CompactSplit 
compactSplitThread;
+275
+276  /**
+277   * Map of regions currently being 
served by this region server. Key is the
+278   * encoded region name.  All access 
should be synchronized.
+279   */
+280  protected final MapString, 
HRegion onlineRegions = new ConcurrentHashMap();
+281
+282  /**
+283   * Map of encoded region names to the 
DataNode locations they should be hosted on
+284   * We store the value as 
InetSocketAddress since this is used only in HDFS
+285   * API (create() that takes favored 
nodes as hints for placing file blocks).
+286   * We could have used ServerName here 
as the value class, but we'd need to
+287   * convert it to InetSocketAddress at 
some point before the HDFS API call, and
+288   * it seems a bit weird to store 
ServerName since ServerName refers to RegionServers
+289   * and here we really mean DataNode 
locations.
+290   */
+291  protected final MapString, 
InetSocketAddress[] regionFavoredNodesMap =
+292  new ConcurrentHashMap();
+293
+294  // Leases
+295  protected Leases leases;
+296
+297  // Instance of the hbase executor 
executorService.
+298  protected ExecutorService 
executorService;
+299
+300  // If false, the file system has become 
unavailable
+301  protected volatile boolean fsOk;
+302  protected HFileSystem fs;
+303  protected HFileSystem walFs;
 304
-305  // Instance of the hbase executor 
executorService.
-306  protected ExecutorService 
executorService;
-307
-308  // If false, the file system has become 
unavailable
-309  protected volatile boolean fsOk;
-310  protected HFileSystem fs;
-311  protected HFileSystem walFs;
-312
-313  // Set when a report to the master 
comes back with a message asking us to
-314  // shutdown. Also set by call to stop 
when debugging or running unit tests
-315  // of HRegionServer in isolation.
-316  private volatile boolean stopped = 
false;
-317
-318  // Go down hard. Used if file system 
becomes unavailable and also in
-319  // debugging and unit tests.
-320  private volatile boolean 
abortRequested;
-321  public 

[26/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.

2018-11-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/procedure/CreateNamespaceProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/CreateNamespaceProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/CreateNamespaceProcedure.html
index 3b44f50..c5e9980 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/CreateNamespaceProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/CreateNamespaceProcedure.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":9,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":9,"i9":10,"i10":9,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":9,"i18":10,"i19":9,"i20":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":9,"i13":10};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -129,7 +129,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class CreateNamespaceProcedure
+public class CreateNamespaceProcedure
 extends AbstractStateMachineNamespaceProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateNamespaceState
 The procedure to create a new namespace.
 
@@ -187,10 +187,6 @@ extends private NamespaceDescriptor
 nsDescriptor
 
-
-private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
-traceEnabled
-
 
 
 
@@ -206,6 +202,13 @@ extends Procedure
 NO_PROC_ID,
 NO_TIMEOUT
 
+
+
+
+
+Fields inherited from 
interfaceorg.apache.hadoop.hbase.master.procedure.TableProcedureInterface
+DUMMY_NAMESPACE_TABLE_NAME
+
 
 
 
@@ -253,124 +256,82 @@ extends 
-protected static void
-createDirectory(MasterProcedureEnvenv,
-   NamespaceDescriptornsDescriptor)
-Create the namespace directory
-
-
-
 protected void
 deserializeStateData(ProcedureStateSerializerserializer)
 Called on store load to allow the user to decode the 
previously serialized
  state.
 
 
-
+
 protected StateMachineProcedure.Flow
 executeFromState(MasterProcedureEnvenv,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateNamespaceStatestate)
 called to perform a single step of the specified 'state' of 
the procedure
 
 
-
+
 protected 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateNamespaceState
 getInitialState()
 Return the initial state object that will be used for the 
first call to executeFromState().
 
 
-
+
 protected https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 getNamespaceName()
 
-
+
 protected 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateNamespaceState
 getState(intstateId)
 Convert an ordinal (or state id) to an Enum (or more 
descriptive) state object.
 
 
-
+
 protected int
 getStateId(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateNamespaceStatestate)
 Convert the Enum (or more descriptive) state object to an 
ordinal (or state id).
 
 
-
-private static TableNamespaceManager
-getTableNamespaceManager(MasterProcedureEnvenv)
-
-
+
 TableProcedureInterface.TableOperationType
 getTableOperationType()
 Given an operation type we can take decisions about what to 
do with pending operations.
 
 
-
-protected static void
-insertIntoNSTable(MasterProcedureEnvenv,
- NamespaceDescriptornsDescriptor)
-Insert the row into ns table
-
-
-
-private boolean
-isBootstrapNamespace()
-
-
+
 protected boolean
 isRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateNamespaceStatestate)
 Used by the default implementation of abort() to know if 
the current state can be aborted
  and rollback can be triggered.
 
 
-
-private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
-isTraceEnabled()
-The procedure could be restarted from a different 
machine.
-
-
-
+
 private boolean
 prepareCreate(MasterProcedureEnvenv)
 Action before any real action of creating namespace.
 
 
-
+
 protected void
 rollbackState(MasterProcedureEnvenv,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CreateNamespaceStatestate)
 called to perform the rollback of the specified state
 
 
-
+
 protected void
 serializeStateData(ProcedureStateSerializerserializer)
 The user-level code of the procedure may have some state to
  persist (e.g.
 
 
-
-protected static void
+
+private static void
 setNamespaceQuota(MasterProcedureEnvenv,
  

[26/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.

2018-11-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionRemoteProcedureBase.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionRemoteProcedureBase.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionRemoteProcedureBase.html
index daf2583..25ef9bf 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionRemoteProcedureBase.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionRemoteProcedureBase.html
@@ -116,75 +116,87 @@
 108  }
 109
 110  @Override
-111  protected void 
rollback(MasterProcedureEnv env) throws IOException, InterruptedException {
-112throw new 
UnsupportedOperationException();
-113  }
-114
-115  @Override
-116  protected boolean 
abort(MasterProcedureEnv env) {
-117return false;
-118  }
-119
-120  /**
-121   * Check whether we still need to make 
the call to RS.
-122   * p/
-123   * Usually this will not happen if we 
do not allow assigning a already onlined region. But if we
-124   * have something wrong in the 
RSProcedureDispatcher, where we have already sent the request to
-125   * RS, but then we tell the upper layer 
the remote call is failed due to rpc timeout or connection
-126   * closed or anything else, then this 
issue can still happen. So here we add a check to make it
-127   * more robust.
-128   */
-129  protected abstract boolean 
shouldDispatch(RegionStateNode regionNode);
+111  protected boolean 
waitInitialized(MasterProcedureEnv env) {
+112if 
(TableName.isMetaTableName(getTableName())) {
+113  return false;
+114}
+115// First we need meta to be loaded, 
and second, if meta is not online then we will likely to
+116// fail when updating meta so we wait 
until it is assigned.
+117AssignmentManager am = 
env.getAssignmentManager();
+118return am.waitMetaLoaded(this) || 
am.waitMetaAssigned(this, region);
+119  }
+120
+121  @Override
+122  protected void 
rollback(MasterProcedureEnv env) throws IOException, InterruptedException {
+123throw new 
UnsupportedOperationException();
+124  }
+125
+126  @Override
+127  protected boolean 
abort(MasterProcedureEnv env) {
+128return false;
+129  }
 130
-131  @Override
-132  protected 
ProcedureMasterProcedureEnv[] execute(MasterProcedureEnv env)
-133  throws ProcedureYieldException, 
ProcedureSuspendedException, InterruptedException {
-134if (dispatched) {
-135  // we are done, the parent 
procedure will check whether we are succeeded.
-136  return null;
-137}
-138RegionStateNode regionNode = 
getRegionNode(env);
-139regionNode.lock();
-140try {
-141  if (!shouldDispatch(regionNode)) 
{
-142return null;
-143  }
-144  // The code which wakes us up also 
needs to lock the RSN so here we do not need to synchronize
-145  // on the event.
-146  ProcedureEvent? event = 
regionNode.getProcedureEvent();
-147  try {
-148
env.getRemoteDispatcher().addOperationToNode(targetServer, this);
-149  } catch 
(FailedRemoteDispatchException e) {
-150LOG.warn("Can not add remote 
operation {} for region {} to server {}, this usually " +
-151  "because the server is alread 
dead, give up and mark the procedure as complete, " +
-152  "the parent procedure will take 
care of this.", this, region, targetServer, e);
-153return null;
-154  }
-155  dispatched = true;
-156  event.suspend();
-157  event.suspendIfNotReady(this);
-158  throw new 
ProcedureSuspendedException();
-159} finally {
-160  regionNode.unlock();
-161}
-162  }
-163
-164  @Override
-165  protected void 
serializeStateData(ProcedureStateSerializer serializer) throws IOException {
-166
serializer.serialize(RegionRemoteProcedureBaseStateData.newBuilder()
-167  
.setRegion(ProtobufUtil.toRegionInfo(region))
-168  
.setTargetServer(ProtobufUtil.toServerName(targetServer)).setDispatched(dispatched).build());
-169  }
-170
-171  @Override
-172  protected void 
deserializeStateData(ProcedureStateSerializer serializer) throws IOException 
{
-173RegionRemoteProcedureBaseStateData 
data =
-174  
serializer.deserialize(RegionRemoteProcedureBaseStateData.class);
-175region = 
ProtobufUtil.toRegionInfo(data.getRegion());
-176targetServer = 
ProtobufUtil.toServerName(data.getTargetServer());
-177dispatched = data.getDispatched();
-178  }
-179}
+131  /**
+132   * Check whether we still need to make 
the call to RS.
+133   * p/
+134   * This could happen when master 
restarts. Since we do not know whether a request has already been
+135   * sent to the region server after we 
add a remote operation to the dispatcher, so the safe way is
+136   * to not persist the dispatched field 
and try to add the remote operation again. But it is
+137   * 

[26/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.

2018-10-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
index 0af8acd..c5f21ac 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html
@@ -645,1615 +645,1597 @@
 637
proc.afterReplay(getEnvironment());
 638  }
 639});
-640
-641// 4. Push the procedures to the 
timeout executor
-642waitingTimeoutList.forEach(proc - 
{
-643  
proc.afterReplay(getEnvironment());
-644  timeoutExecutor.add(proc);
-645});
-646// 5. restore locks
-647restoreLocks();
-648// 6. Push the procedure to the 
scheduler
-649
failedList.forEach(scheduler::addBack);
-650runnableList.forEach(p - {
-651  p.afterReplay(getEnvironment());
-652  if (!p.hasParent()) {
-653
sendProcedureLoadedNotification(p.getProcId());
-654  }
-655  // If the procedure holds the lock, 
put the procedure in front
-656  // If its parent holds the lock, 
put the procedure in front
-657  // TODO. Is that possible that its 
ancestor holds the lock?
-658  // For now, the deepest procedure 
hierarchy is:
-659  // ModifyTableProcedure - 
ReopenTableProcedure -
-660  // MoveTableProcedure - 
Unassign/AssignProcedure
-661  // But ModifyTableProcedure and 
ReopenTableProcedure won't hold the lock
-662  // So, check parent lock is 
enough(a tricky case is resovled by HBASE-21384).
-663  // If some one change or add new 
procedures making 'grandpa' procedure
-664  // holds the lock, but parent 
procedure don't hold the lock, there will
-665  // be a problem here. We have to 
check one procedure's ancestors.
-666  // And we need to change 
LockAndQueue.hasParentLock(Procedure? proc) method
-667  // to check all ancestors too.
-668  if (p.isLockedWhenLoading() || 
(p.hasParent()  procedures
-669  
.get(p.getParentProcId()).isLockedWhenLoading())) {
-670scheduler.addFront(p, false);
-671  } else {
-672// if it was not, it can wait.
-673scheduler.addBack(p, false);
-674  }
-675});
-676// After all procedures put into the 
queue, signal the worker threads.
-677// Otherwise, there is a race 
condition. See HBASE-21364.
-678scheduler.signalAll();
-679  }
+640// 4. restore locks
+641restoreLocks();
+642
+643// 5. Push the procedures to the 
timeout executor
+644waitingTimeoutList.forEach(proc - 
{
+645  
proc.afterReplay(getEnvironment());
+646  timeoutExecutor.add(proc);
+647});
+648
+649// 6. Push the procedure to the 
scheduler
+650
failedList.forEach(scheduler::addBack);
+651runnableList.forEach(p - {
+652  p.afterReplay(getEnvironment());
+653  if (!p.hasParent()) {
+654
sendProcedureLoadedNotification(p.getProcId());
+655  }
+656  scheduler.addBack(p);
+657});
+658// After all procedures put into the 
queue, signal the worker threads.
+659// Otherwise, there is a race 
condition. See HBASE-21364.
+660scheduler.signalAll();
+661  }
+662
+663  /**
+664   * Initialize the procedure executor, 
but do not start workers. We will start them later.
+665   * p/
+666   * It calls 
ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, 
and
+667   * ensure a single executor, and start 
the procedure replay to resume and recover the previous
+668   * pending and in-progress 
procedures.
+669   * @param numThreads number of threads 
available for procedure execution.
+670   * @param abortOnCorruption true if you 
want to abort your service in case a corrupted procedure
+671   *  is found on replay. 
otherwise false.
+672   */
+673  public void init(int numThreads, 
boolean abortOnCorruption) throws IOException {
+674// We have numThreads executor + one 
timer thread used for timing out
+675// procedures and triggering periodic 
procedures.
+676this.corePoolSize = numThreads;
+677this.maxPoolSize = 10 * numThreads;
+678LOG.info("Starting {} core workers 
(bigger of cpus/4 or 16) with max (burst) worker count={}",
+679corePoolSize, maxPoolSize);
 680
-681  /**
-682   * Initialize the procedure executor, 
but do not start workers. We will start them later.
-683   * p/
-684   * It calls 
ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, 
and
-685   * ensure a single executor, and start 
the procedure replay to resume and recover the previous
-686   * pending and in-progress 
procedures.
-687   * @param numThreads number of threads 

[26/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.

2018-10-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
index c7d99b2..9d1542c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
@@ -382,1357 +382,1365 @@
 374for (int i = 0; i  
this.curFunctionCosts.length; i++) {
 375  curFunctionCosts[i] = 
tempFunctionCosts[i];
 376}
-377LOG.info("start 
StochasticLoadBalancer.balancer, initCost=" + currentCost + ", functionCost="
-378+ functionCost());
+377double initCost = currentCost;
+378double newCost = currentCost;
 379
-380double initCost = currentCost;
-381double newCost = currentCost;
-382
-383long computedMaxSteps;
-384if (runMaxSteps) {
-385  computedMaxSteps = 
Math.max(this.maxSteps,
-386  ((long)cluster.numRegions * 
(long)this.stepsPerRegion * (long)cluster.numServers));
-387} else {
-388  computedMaxSteps = 
Math.min(this.maxSteps,
-389  ((long)cluster.numRegions * 
(long)this.stepsPerRegion * (long)cluster.numServers));
-390}
-391// Perform a stochastic walk to see 
if we can get a good fit.
-392long step;
-393
-394for (step = 0; step  
computedMaxSteps; step++) {
-395  Cluster.Action action = 
nextAction(cluster);
-396
-397  if (action.type == Type.NULL) {
-398continue;
-399  }
-400
-401  cluster.doAction(action);
-402  updateCostsWithAction(cluster, 
action);
-403
-404  newCost = computeCost(cluster, 
currentCost);
-405
-406  // Should this be kept?
-407  if (newCost  currentCost) {
-408currentCost = newCost;
-409
-410// save for JMX
-411curOverallCost = currentCost;
-412for (int i = 0; i  
this.curFunctionCosts.length; i++) {
-413  curFunctionCosts[i] = 
tempFunctionCosts[i];
-414}
-415  } else {
-416// Put things back the way they 
were before.
-417// TODO: undo by remembering old 
values
-418Action undoAction = 
action.undoAction();
-419cluster.doAction(undoAction);
-420updateCostsWithAction(cluster, 
undoAction);
-421  }
-422
-423  if 
(EnvironmentEdgeManager.currentTime() - startTime 
-424  maxRunningTime) {
-425break;
-426  }
-427}
-428long endTime = 
EnvironmentEdgeManager.currentTime();
-429
-430
metricsBalancer.balanceCluster(endTime - startTime);
-431
-432// update costs metrics
-433updateStochasticCosts(tableName, 
curOverallCost, curFunctionCosts);
-434if (initCost  currentCost) {
-435  plans = 
createRegionPlans(cluster);
-436  LOG.info("Finished computing new 
load balance plan. Computation took {}" +
-437" to try {} different iterations. 
 Found a solution that moves " +
-438"{} regions; Going from a 
computed cost of {}" +
-439" to a new cost of {}", 
java.time.Duration.ofMillis(endTime - startTime),
-440step, plans.size(), initCost, 
currentCost);
-441  return plans;
-442}
-443LOG.info("Could not find a better 
load balance plan.  Tried {} different configurations in " +
-444  "{}, and did not find anything with 
a computed cost less than {}", step,
-445  java.time.Duration.ofMillis(endTime 
- startTime), initCost);
-446return null;
-447  }
-448
-449  /**
-450   * update costs to JMX
-451   */
-452  private void 
updateStochasticCosts(TableName tableName, Double overall, Double[] subCosts) 
{
-453if (tableName == null) return;
-454
-455// check if the metricsBalancer is 
MetricsStochasticBalancer before casting
-456if (metricsBalancer instanceof 
MetricsStochasticBalancer) {
-457  MetricsStochasticBalancer balancer 
= (MetricsStochasticBalancer) metricsBalancer;
-458  // overall cost
-459  
balancer.updateStochasticCost(tableName.getNameAsString(),
-460"Overall", "Overall cost", 
overall);
-461
-462  // each cost function
-463  for (int i = 0; i  
costFunctions.length; i++) {
-464CostFunction costFunction = 
costFunctions[i];
-465String costFunctionName = 
costFunction.getClass().getSimpleName();
-466Double costPercent = (overall == 
0) ? 0 : (subCosts[i] / overall);
-467// TODO: cost function may need a 
specific description
-468
balancer.updateStochasticCost(tableName.getNameAsString(), costFunctionName,
-469  "The percent of " + 

[26/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.

2018-10-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.html
new file mode 100644
index 000..8ef0ff5
--- /dev/null
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.html
@@ -0,0 +1,371 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/**
+002 * Licensed to the Apache Software 
Foundation (ASF) under one
+003 * or more contributor license 
agreements.  See the NOTICE file
+004 * distributed with this work for 
additional information
+005 * regarding copyright ownership.  The 
ASF licenses this file
+006 * to you under the Apache License, 
Version 2.0 (the
+007 * "License"); you may not use this file 
except in compliance
+008 * with the License.  You may obtain a 
copy of the License at
+009 *
+010 * 
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or 
agreed to in writing, software
+013 * distributed under the License is 
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+015 * See the License for the specific 
language governing permissions and
+016 * limitations under the License.
+017 */
+018package 
org.apache.hadoop.hbase.procedure2.store.wal;
+019
+020import java.io.IOException;
+021import java.util.ArrayList;
+022import java.util.Collection;
+023import java.util.Collections;
+024import java.util.Comparator;
+025import java.util.HashMap;
+026import java.util.Iterator;
+027import java.util.List;
+028import java.util.Map;
+029import 
java.util.NoSuchElementException;
+030import 
org.apache.commons.lang3.mutable.MutableInt;
+031import 
org.apache.hadoop.hbase.procedure2.Procedure;
+032import 
org.apache.hadoop.hbase.procedure2.ProcedureUtil;
+033import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
+034import 
org.apache.yetus.audience.InterfaceAudience;
+035import org.slf4j.Logger;
+036import org.slf4j.LoggerFactory;
+037
+038import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos;
+039
+040/**
+041 * Used to build the tree for 
procedures.
+042 * p/
+043 * We will group the procedures with the 
root procedure, and then validate each group. For each
+044 * group of procedures(with the same root 
procedure), we will collect all the stack ids, if the max
+045 * stack id is n, then all the stack ids 
should be from 0 to n, non-repetition and non-omission. If
+046 * not, we will consider all the 
procedures in this group as corrupted. Please see the code in
+047 * {@link #checkReady(Entry, Map)} 
method.
+048 * p/
+049 * For the procedures not in any group, 
i.e, can not find the root procedure for these procedures,
+050 * we will also consider them as 
corrupted. Please see the code in {@link #checkOrphan(Map)} method.
+051 */
+052@InterfaceAudience.Private
+053public final class WALProcedureTree {
+054
+055  private static final Logger LOG = 
LoggerFactory.getLogger(WALProcedureTree.class);
+056
+057  private static final class Entry {
+058
+059private final 
ProcedureProtos.Procedure proc;
+060
+061private final ListEntry 
subProcs = new ArrayList();
+062
+063public 
Entry(ProcedureProtos.Procedure proc) {
+064  this.proc = proc;
+065}
+066
+067@Override
+068public String toString() {
+069  StringBuilder sb = new 
StringBuilder();
+070  sb.append("Procedure(pid=");
+071  sb.append(proc.getProcId());
+072  sb.append(", ppid=");
+073  sb.append(proc.hasParentId() ? 
proc.getParentId() : Procedure.NO_PROC_ID);
+074  sb.append(", class=");
+075  sb.append(proc.getClassName());
+076  sb.append(")");
+077  return sb.toString();
+078}
+079  }
+080
+081  // when loading we will iterator the 
procedures twice, so use this class to cache the deserialized
+082  // result to prevent deserializing 
multiple times.
+083  private static final class ProtoAndProc 
{
+084private final 
ProcedureProtos.Procedure proto;
+085
+086private Procedure? proc;
+087
+088public 
ProtoAndProc(ProcedureProtos.Procedure proto) {
+089  this.proto = proto;
+090}
+091
+092public Procedure? getProc() 
throws IOException {
+093  if (proc == null) {
+094proc = 
ProcedureUtil.convertToProcedure(proto);
+095  }
+096  return proc;
+097}
+098  }
+099
+100  private final ListProtoAndProc 
validProcs = new ArrayList();
+101
+102  private final ListProtoAndProc 
corruptedProcs = new ArrayList();
+103
+104  private static boolean 
isFinished(ProcedureProtos.Procedure proc) {
+105if (!proc.hasParentId()) {
+106  switch (proc.getState()) {
+107case 

[26/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.

2018-10-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineNamespaceProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineNamespaceProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineNamespaceProcedure.html
index 45338a8..6ca7efe 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineNamespaceProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineNamespaceProcedure.html
@@ -299,7 +299,7 @@ implements Procedure
-addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doExecute,
 doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics,
 getProcId,
 getProcIdHashCode, getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime,
 getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 holdLock, 
incChildrenLatch,
 isBypass,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex, setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner, setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime, setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
 skipPersistence,
 toString,
 toStringClass,
 toStringDetails,
 to
 StringSimpleSB, updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
+addStackIndex,
 afterReplay,
 beforeReplay,
 bypass,
 compareTo,
 completionCleanup,
 doExecute,
 doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics,
 getProcId, getProcIdHashCode,
 getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime,
 <
 a 
href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#getTimeout--">getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent, holdLock,
 incChildrenLatch,
 isBypass,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting, removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey, setOwner,
 setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState, setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
 skipPersistence,
 toString,
 toStringClass,
 toStringD
 etails, toStringSimpleSB,
 updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.html
index d014986..9f072ec 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineRegionProcedure.html
@@ -331,7 +331,7 @@ extends Procedure
-addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doExecute,
 doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics,
 getProcId,
 getProcIdHashCode, getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime,
 getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 incChildrenLatch,
 isBypass,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner,
 setOwner, setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout, setTimeoutFailure,
 shouldWaitClientAck,
 skipPersistence,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB,
 updateMetricsOnFinish, updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
+addStackIndex,
 afterReplay,
 beforeReplay,
 bypass,
 compareTo,
 completionCleanup,
 doExecute,
 doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics,
 getProcId, 

[26/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.

2018-10-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/client/RestoreSnapshotFromClientAfterSplittingRegionsTestBase.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/client/RestoreSnapshotFromClientAfterSplittingRegionsTestBase.html
 
b/testdevapidocs/org/apache/hadoop/hbase/client/RestoreSnapshotFromClientAfterSplittingRegionsTestBase.html
new file mode 100644
index 000..5559c32
--- /dev/null
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/client/RestoreSnapshotFromClientAfterSplittingRegionsTestBase.html
@@ -0,0 +1,310 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+RestoreSnapshotFromClientAfterSplittingRegionsTestBase (Apache HBase 
3.0.0-SNAPSHOT Test API)
+
+
+
+
+
+var methods = {"i0":10};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.client
+Class RestoreSnapshotFromClientAfterSplittingRegionsTestBase
+
+
+
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase
+
+
+org.apache.hadoop.hbase.client.RestoreSnapshotFromClientAfterSplittingRegionsTestBase
+
+
+
+
+
+
+
+
+
+Direct Known Subclasses:
+TestMobRestoreSnapshotFromClientAfterSplittingRegions,
 TestRestoreSnapshotFromClientAfterSplittingRegions
+
+
+
+public class RestoreSnapshotFromClientAfterSplittingRegionsTestBase
+extends RestoreSnapshotFromClientTestBase
+
+
+
+
+
+
+
+
+
+
+
+Field Summary
+
+
+
+
+Fields inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase
+admin,
 emptySnapshot,
 FAMILY,
 name,
 snapshot0Rows,
 snapshot1Rows,
 snapshotName0,
 snapshotName1,
 snapshotName2,
 tableName,
 TEST_FAMILY2,
 TEST_UTIL
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+RestoreSnapshotFromClientAfterSplittingRegionsTestBase()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsInstance MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+void
+testRestoreSnapshotAfterSplittingRegions()
+
+
+
+
+
+
+Methods inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase
+countRows,
 createTable,
 getNumReplicas,
 getValidMethodName,
 setup,
 setupCluster,
 setupConf,
 splitRegion,
 tearDown,
 tearDownAfterClass,
 verifyRowCount
+
+
+
+
+
+Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
 title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
 title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
 title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/ja
 va/lang/Object.html?is-external=true#notifyAll--" title="class or interface in 
java.lang">notifyAll, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
 title="class or interface in java.lang">toString, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--;
 title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-;
 

[26/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.

2018-10-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
index 061ce80..bdfc3f8 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
@@ -39,2126 +39,2163 @@
 031import java.util.Set;
 032import 
java.util.concurrent.ConcurrentHashMap;
 033import 
java.util.concurrent.CopyOnWriteArrayList;
-034import java.util.concurrent.TimeUnit;
-035import 
java.util.concurrent.atomic.AtomicBoolean;
-036import 
java.util.concurrent.atomic.AtomicInteger;
-037import 
java.util.concurrent.atomic.AtomicLong;
-038import java.util.stream.Collectors;
-039import java.util.stream.Stream;
-040import 
org.apache.hadoop.conf.Configuration;
-041import 
org.apache.hadoop.hbase.HConstants;
-042import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
-043import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-044import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
-045import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
-046import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
-047import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
-048import 
org.apache.hadoop.hbase.security.User;
-049import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-050import 
org.apache.hadoop.hbase.util.IdLock;
-051import 
org.apache.hadoop.hbase.util.NonceKey;
-052import 
org.apache.hadoop.hbase.util.Threads;
-053import 
org.apache.yetus.audience.InterfaceAudience;
-054import org.slf4j.Logger;
-055import org.slf4j.LoggerFactory;
-056
-057import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-058import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+034import java.util.concurrent.Executor;
+035import java.util.concurrent.Executors;
+036import java.util.concurrent.TimeUnit;
+037import 
java.util.concurrent.atomic.AtomicBoolean;
+038import 
java.util.concurrent.atomic.AtomicInteger;
+039import 
java.util.concurrent.atomic.AtomicLong;
+040import java.util.stream.Collectors;
+041import java.util.stream.Stream;
+042import 
org.apache.hadoop.conf.Configuration;
+043import 
org.apache.hadoop.hbase.HConstants;
+044import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
+045import 
org.apache.hadoop.hbase.log.HBaseMarkers;
+046import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
+047import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
+048import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
+049import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureStoreListener;
+050import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
+051import 
org.apache.hadoop.hbase.security.User;
+052import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+053import 
org.apache.hadoop.hbase.util.IdLock;
+054import 
org.apache.hadoop.hbase.util.NonceKey;
+055import 
org.apache.hadoop.hbase.util.Threads;
+056import 
org.apache.yetus.audience.InterfaceAudience;
+057import org.slf4j.Logger;
+058import org.slf4j.LoggerFactory;
 059
-060import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
-061
-062/**
-063 * Thread Pool that executes the 
submitted procedures.
-064 * The executor has a ProcedureStore 
associated.
-065 * Each operation is logged and on 
restart the pending procedures are resumed.
-066 *
-067 * Unless the Procedure code throws an 
error (e.g. invalid user input)
-068 * the procedure will complete (at some 
point in time), On restart the pending
-069 * procedures are resumed and the once 
failed will be rolledback.
+060import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+061import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+062import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
+063
+064import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
+065
+066/**
+067 * Thread Pool that executes the 
submitted procedures.
+068 * The executor has a ProcedureStore 
associated.
+069 * Each operation is logged and on 
restart the pending procedures are resumed.
 070 *
-071 * The user can add procedures to the 
executor via submitProcedure(proc)
-072 * check for the finished state via 
isFinished(procId)
-073 * and get the result via 
getResult(procId)
-074 */
-075@InterfaceAudience.Private
-076public class 
ProcedureExecutorTEnvironment {
-077  private static final Logger LOG = 
LoggerFactory.getLogger(ProcedureExecutor.class);

[26/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.

2018-09-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
index 43c66a8..061ce80 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
@@ -23,2136 +23,2142 @@
 015 * See the License for the specific 
language governing permissions and
 016 * limitations under the License.
 017 */
-018
-019package 
org.apache.hadoop.hbase.procedure2;
-020
-021import java.io.IOException;
-022import java.util.ArrayDeque;
-023import java.util.ArrayList;
-024import java.util.Arrays;
-025import java.util.Collection;
-026import java.util.Deque;
-027import java.util.HashSet;
-028import java.util.Iterator;
-029import java.util.List;
-030import java.util.Map;
-031import java.util.Objects;
-032import java.util.Set;
-033import 
java.util.concurrent.ConcurrentHashMap;
-034import 
java.util.concurrent.CopyOnWriteArrayList;
-035import java.util.concurrent.TimeUnit;
-036import 
java.util.concurrent.atomic.AtomicBoolean;
-037import 
java.util.concurrent.atomic.AtomicInteger;
-038import 
java.util.concurrent.atomic.AtomicLong;
-039import java.util.stream.Collectors;
-040import java.util.stream.Stream;
-041
-042import 
org.apache.hadoop.conf.Configuration;
-043import 
org.apache.hadoop.hbase.HConstants;
-044import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
-045import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-046import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
-047import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
-048import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
-049import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
-050import 
org.apache.hadoop.hbase.security.User;
-051import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-052import 
org.apache.hadoop.hbase.util.IdLock;
-053import 
org.apache.hadoop.hbase.util.NonceKey;
-054import 
org.apache.hadoop.hbase.util.Threads;
-055import 
org.apache.yetus.audience.InterfaceAudience;
-056import org.slf4j.Logger;
-057import org.slf4j.LoggerFactory;
-058
-059import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-060import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+018package 
org.apache.hadoop.hbase.procedure2;
+019
+020import java.io.IOException;
+021import java.util.ArrayDeque;
+022import java.util.ArrayList;
+023import java.util.Arrays;
+024import java.util.Collection;
+025import java.util.Deque;
+026import java.util.HashSet;
+027import java.util.Iterator;
+028import java.util.List;
+029import java.util.Map;
+030import java.util.Objects;
+031import java.util.Set;
+032import 
java.util.concurrent.ConcurrentHashMap;
+033import 
java.util.concurrent.CopyOnWriteArrayList;
+034import java.util.concurrent.TimeUnit;
+035import 
java.util.concurrent.atomic.AtomicBoolean;
+036import 
java.util.concurrent.atomic.AtomicInteger;
+037import 
java.util.concurrent.atomic.AtomicLong;
+038import java.util.stream.Collectors;
+039import java.util.stream.Stream;
+040import 
org.apache.hadoop.conf.Configuration;
+041import 
org.apache.hadoop.hbase.HConstants;
+042import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
+043import 
org.apache.hadoop.hbase.log.HBaseMarkers;
+044import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
+045import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
+046import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
+047import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
+048import 
org.apache.hadoop.hbase.security.User;
+049import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+050import 
org.apache.hadoop.hbase.util.IdLock;
+051import 
org.apache.hadoop.hbase.util.NonceKey;
+052import 
org.apache.hadoop.hbase.util.Threads;
+053import 
org.apache.yetus.audience.InterfaceAudience;
+054import org.slf4j.Logger;
+055import org.slf4j.LoggerFactory;
+056
+057import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+058import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+059
+060import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
 061
-062import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
-063
-064/**
-065 * Thread Pool that executes the 
submitted procedures.
-066 * The executor has a ProcedureStore 
associated.
-067 * Each operation is logged and on 
restart the 

[26/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.

2018-09-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html 
b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html
index e3153b8..2651660 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.DfsBuilderUtility.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class CommonFSUtils.DfsBuilderUtility
+private static class CommonFSUtils.DfsBuilderUtility
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 
 
@@ -257,7 +257,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 dfsClass
-statichttps://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? dfsClass
+statichttps://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? dfsClass
 
 
 
@@ -266,7 +266,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 createMethod
-statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method createMethod
+statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method createMethod
 
 
 
@@ -275,7 +275,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 overwriteMethod
-statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method overwriteMethod
+statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method overwriteMethod
 
 
 
@@ -284,7 +284,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 bufferSizeMethod
-statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method bufferSizeMethod
+statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method bufferSizeMethod
 
 
 
@@ -293,7 +293,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 blockSizeMethod
-statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method blockSizeMethod
+statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method blockSizeMethod
 
 
 
@@ -302,7 +302,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 recursiveMethod
-statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method recursiveMethod
+statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method recursiveMethod
 
 
 
@@ -311,7 +311,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 replicateMethod
-statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method replicateMethod
+statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method replicateMethod
 
 
 
@@ -320,7 +320,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 replicationMethod
-statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method replicationMethod
+statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method replicationMethod
 
 
 
@@ -329,7 +329,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 buildMethod
-statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method buildMethod
+statichttps://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Method buildMethod
 
 
 
@@ -338,7 +338,7 @@ extends 

[26/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.

2018-09-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.TrackingThreadPoolExecutor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.TrackingThreadPoolExecutor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.TrackingThreadPoolExecutor.html
index 49f081b..33c9cc0 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.TrackingThreadPoolExecutor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.TrackingThreadPoolExecutor.html
@@ -35,309 +35,328 @@
 027import 
java.util.concurrent.BlockingQueue;
 028import 
java.util.concurrent.ConcurrentHashMap;
 029import 
java.util.concurrent.ConcurrentMap;
-030import 
java.util.concurrent.LinkedBlockingQueue;
-031import 
java.util.concurrent.ThreadPoolExecutor;
-032import java.util.concurrent.TimeUnit;
-033import 
java.util.concurrent.atomic.AtomicLong;
-034
-035import 
org.apache.yetus.audience.InterfaceAudience;
-036import org.slf4j.Logger;
-037import org.slf4j.LoggerFactory;
-038import 
org.apache.hadoop.hbase.monitoring.ThreadMonitoring;
+030import java.util.concurrent.Executors;
+031import 
java.util.concurrent.LinkedBlockingQueue;
+032import 
java.util.concurrent.ThreadPoolExecutor;
+033import java.util.concurrent.TimeUnit;
+034import 
java.util.concurrent.atomic.AtomicLong;
+035import 
org.apache.hadoop.hbase.monitoring.ThreadMonitoring;
+036import 
org.apache.yetus.audience.InterfaceAudience;
+037import org.slf4j.Logger;
+038import org.slf4j.LoggerFactory;
 039
 040import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
 041import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
 042import 
org.apache.hbase.thirdparty.com.google.common.collect.Maps;
-043import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-044
-045/**
-046 * This is a generic executor service. 
This component abstracts a
-047 * threadpool, a queue to which {@link 
EventType}s can be submitted,
-048 * and a 
codeRunnable/code that handles the object that is added to the 
queue.
-049 *
-050 * pIn order to create a new 
service, create an instance of this class and
-051 * then do: 
codeinstance.startExecutorService("myService");/code.  When 
done
-052 * call {@link #shutdown()}.
-053 *
-054 * pIn order to use the service 
created above, call
-055 * {@link #submit(EventHandler)}.
-056 */
-057@InterfaceAudience.Private
-058public class ExecutorService {
-059  private static final Logger LOG = 
LoggerFactory.getLogger(ExecutorService.class);
-060
-061  // hold the all the executors created 
in a map addressable by their names
-062  private final 
ConcurrentHashMapString, Executor executorMap = new 
ConcurrentHashMap();
+043import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ListenableFuture;
+044import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ListeningScheduledExecutorService;
+045import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.MoreExecutors;
+046import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
+047
+048/**
+049 * This is a generic executor service. 
This component abstracts a
+050 * threadpool, a queue to which {@link 
EventType}s can be submitted,
+051 * and a 
codeRunnable/code that handles the object that is added to the 
queue.
+052 *
+053 * pIn order to create a new 
service, create an instance of this class and
+054 * then do: 
codeinstance.startExecutorService("myService");/code.  When 
done
+055 * call {@link #shutdown()}.
+056 *
+057 * pIn order to use the service 
created above, call
+058 * {@link #submit(EventHandler)}.
+059 */
+060@InterfaceAudience.Private
+061public class ExecutorService {
+062  private static final Logger LOG = 
LoggerFactory.getLogger(ExecutorService.class);
 063
-064  // Name of the server hosting this 
executor service.
-065  private final String servername;
+064  // hold the all the executors created 
in a map addressable by their names
+065  private final 
ConcurrentHashMapString, Executor executorMap = new 
ConcurrentHashMap();
 066
-067  /**
-068   * Default constructor.
-069   * @param servername Name of the 
hosting server.
-070   */
-071  public ExecutorService(final String 
servername) {
-072super();
-073this.servername = servername;
-074  }
-075
-076  /**
-077   * Start an executor service with a 
given name. If there was a service already
-078   * started with the same name, this 
throws a RuntimeException.
-079   * @param name Name of the service to 
start.
-080   */
-081  @VisibleForTesting
-082  public void startExecutorService(String 
name, int maxThreads) {
-083if (this.executorMap.get(name) != 
null) {
-084  throw new RuntimeException("An 
executor service with the name " + name +

[26/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.

2018-09-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
index a5789e0..93a57cb 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
@@ -238,4120 +238,4119 @@
 230 * @see Admin
 231 */
 232@InterfaceAudience.Private
-233@InterfaceStability.Evolving
-234public class HBaseAdmin implements Admin 
{
-235  private static final Logger LOG = 
LoggerFactory.getLogger(HBaseAdmin.class);
-236
-237  private ClusterConnection connection;
-238
-239  private final Configuration conf;
-240  private final long pause;
-241  private final int numRetries;
-242  private final int syncWaitTimeout;
-243  private boolean aborted;
-244  private int operationTimeout;
-245  private int rpcTimeout;
-246
-247  private RpcRetryingCallerFactory 
rpcCallerFactory;
-248  private RpcControllerFactory 
rpcControllerFactory;
-249
-250  private NonceGenerator ng;
-251
-252  @Override
-253  public int getOperationTimeout() {
-254return operationTimeout;
-255  }
-256
-257  HBaseAdmin(ClusterConnection 
connection) throws IOException {
-258this.conf = 
connection.getConfiguration();
-259this.connection = connection;
-260
-261// TODO: receive 
ConnectionConfiguration here rather than re-parsing these configs every time.
-262this.pause = 
this.conf.getLong(HConstants.HBASE_CLIENT_PAUSE,
-263
HConstants.DEFAULT_HBASE_CLIENT_PAUSE);
-264this.numRetries = 
this.conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER,
-265
HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER);
-266this.operationTimeout = 
this.conf.getInt(HConstants.HBASE_CLIENT_OPERATION_TIMEOUT,
-267
HConstants.DEFAULT_HBASE_CLIENT_OPERATION_TIMEOUT);
-268this.rpcTimeout = 
this.conf.getInt(HConstants.HBASE_RPC_TIMEOUT_KEY,
-269
HConstants.DEFAULT_HBASE_RPC_TIMEOUT);
-270this.syncWaitTimeout = 
this.conf.getInt(
-271  
"hbase.client.sync.wait.timeout.msec", 10 * 6); // 10min
-272
-273this.rpcCallerFactory = 
connection.getRpcRetryingCallerFactory();
-274this.rpcControllerFactory = 
connection.getRpcControllerFactory();
-275
-276this.ng = 
this.connection.getNonceGenerator();
-277  }
-278
-279  @Override
-280  public void abort(String why, Throwable 
e) {
-281// Currently does nothing but throw 
the passed message and exception
-282this.aborted = true;
-283throw new RuntimeException(why, e);
-284  }
-285
-286  @Override
-287  public boolean isAborted() {
-288return this.aborted;
-289  }
-290
-291  @Override
-292  public boolean abortProcedure(final 
long procId, final boolean mayInterruptIfRunning)
-293  throws IOException {
-294return 
get(abortProcedureAsync(procId, mayInterruptIfRunning), this.syncWaitTimeout,
-295  TimeUnit.MILLISECONDS);
-296  }
-297
-298  @Override
-299  public FutureBoolean 
abortProcedureAsync(final long procId, final boolean mayInterruptIfRunning)
-300  throws IOException {
-301Boolean abortProcResponse =
-302executeCallable(new 
MasterCallableAbortProcedureResponse(getConnection(),
-303getRpcControllerFactory()) 
{
-304  @Override
-305  protected AbortProcedureResponse 
rpcCall() throws Exception {
-306AbortProcedureRequest 
abortProcRequest =
-307
AbortProcedureRequest.newBuilder().setProcId(procId).build();
-308return 
master.abortProcedure(getRpcController(), abortProcRequest);
-309  }
-310}).getIsProcedureAborted();
-311return new AbortProcedureFuture(this, 
procId, abortProcResponse);
-312  }
-313
-314  @Override
-315  public ListTableDescriptor 
listTableDescriptors() throws IOException {
-316return 
listTableDescriptors((Pattern)null, false);
-317  }
-318
-319  @Override
-320  public ListTableDescriptor 
listTableDescriptors(Pattern pattern) throws IOException {
-321return listTableDescriptors(pattern, 
false);
-322  }
-323
-324  @Override
-325  public ListTableDescriptor 
listTableDescriptors(Pattern pattern, boolean includeSysTables)
-326  throws IOException {
-327return executeCallable(new 
MasterCallableListTableDescriptor(getConnection(),
-328getRpcControllerFactory()) {
-329  @Override
-330  protected 
ListTableDescriptor rpcCall() throws Exception {
-331GetTableDescriptorsRequest req 
=
-332
RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables);
-333return 
ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(),
-334req));
-335  }
-336});
-337  }
-338
-339  

[26/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.

2018-09-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyByteBufferExtendedCell.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyByteBufferExtendedCell.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyByteBufferExtendedCell.html
index 1e6a2bb..9881003 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyByteBufferExtendedCell.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyByteBufferExtendedCell.html
@@ -145,332 +145,328 @@
 137
 138  @Override
 139  public boolean equals(Object obj) {
-140if (obj == null || (!(obj instanceof 
KeyOnlyFilter))) {
-141  return false;
-142}
-143KeyOnlyFilter f = (KeyOnlyFilter) 
obj;
-144return 
this.areSerializedFieldsEqual(f);
-145  }
-146
-147  @Override
-148  public int hashCode() {
-149return Objects.hash(this.lenAsVal);
-150  }
+140return obj instanceof Filter 
 areSerializedFieldsEqual((Filter) obj);
+141  }
+142
+143  @Override
+144  public int hashCode() {
+145return Objects.hash(this.lenAsVal);
+146  }
+147
+148  static class KeyOnlyCell implements 
Cell {
+149private Cell cell;
+150private boolean lenAsVal;
 151
-152  static class KeyOnlyCell implements 
Cell {
-153private Cell cell;
-154private boolean lenAsVal;
-155
-156public KeyOnlyCell(Cell c, boolean 
lenAsVal) {
-157  this.cell = c;
-158  this.lenAsVal = lenAsVal;
-159}
-160
-161@Override
-162public byte[] getRowArray() {
-163  return cell.getRowArray();
-164}
-165
-166@Override
-167public int getRowOffset() {
-168  return cell.getRowOffset();
-169}
-170
-171@Override
-172public short getRowLength() {
-173  return cell.getRowLength();
-174}
-175
-176@Override
-177public byte[] getFamilyArray() {
-178  return cell.getFamilyArray();
-179}
-180
-181@Override
-182public int getFamilyOffset() {
-183  return cell.getFamilyOffset();
-184}
-185
-186@Override
-187public byte getFamilyLength() {
-188  return cell.getFamilyLength();
-189}
-190
-191@Override
-192public byte[] getQualifierArray() {
-193  return cell.getQualifierArray();
-194}
-195
-196@Override
-197public int getQualifierOffset() {
-198  return cell.getQualifierOffset();
-199}
-200
-201@Override
-202public int getQualifierLength() {
-203  return cell.getQualifierLength();
-204}
-205
-206@Override
-207public long getTimestamp() {
-208  return cell.getTimestamp();
-209}
-210
-211@Override
-212public byte getTypeByte() {
-213  return cell.getTypeByte();
-214}
-215
-216@Override
-217public Type getType() {
-218  return cell.getType();
-219}
-220
-221
-222@Override
-223public long getSequenceId() {
-224  return 0;
-225}
-226
-227@Override
-228public byte[] getValueArray() {
-229  if (lenAsVal) {
-230return 
Bytes.toBytes(cell.getValueLength());
-231  } else {
-232return 
HConstants.EMPTY_BYTE_ARRAY;
-233  }
-234}
-235
-236@Override
-237public int getValueOffset() {
-238  return 0;
-239}
-240
-241@Override
-242public int getValueLength() {
-243  if (lenAsVal) {
-244return Bytes.SIZEOF_INT;
-245  } else {
-246return 0;
-247  }
-248}
-249
-250@Override
-251public byte[] getTagsArray() {
-252  return 
HConstants.EMPTY_BYTE_ARRAY;
-253}
-254
-255@Override
-256public int getTagsOffset() {
-257  return 0;
-258}
-259
-260@Override
-261public int getTagsLength() {
-262  return 0;
-263}
-264  }
-265
-266  static class 
KeyOnlyByteBufferExtendedCell extends ByteBufferExtendedCell {
-267public static final int 
FIXED_OVERHEAD = ClassSize.OBJECT + ClassSize.REFERENCE
-268+ Bytes.SIZEOF_BOOLEAN;
-269private ByteBufferExtendedCell 
cell;
-270private boolean lenAsVal;
-271
-272public 
KeyOnlyByteBufferExtendedCell(ByteBufferExtendedCell c, boolean lenAsVal) {
-273  this.cell = c;
-274  this.lenAsVal = lenAsVal;
-275}
-276
-277@Override
-278public byte[] getRowArray() {
-279  return cell.getRowArray();
-280}
-281
-282@Override
-283public int getRowOffset() {
-284  return cell.getRowOffset();
-285}
-286
-287@Override
-288public short getRowLength() {
-289  return cell.getRowLength();
-290}
-291
-292@Override
-293public byte[] getFamilyArray() {
-294  return cell.getFamilyArray();
-295}
-296
-297@Override
-298public int getFamilyOffset() {
-299  return cell.getFamilyOffset();
-300}
-301
-302@Override
-303public byte getFamilyLength() {
-304  return 

[26/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.

2018-08-31 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
index cd9aaf5..4f72101 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
@@ -1749,7 +1749,7 @@ implements 
 
 FIXED_OVERHEAD
-public static finallong FIXED_OVERHEAD
+public static finallong FIXED_OVERHEAD
 
 
 
@@ -1758,7 +1758,7 @@ implements 
 
 DEEP_OVERHEAD
-public static finallong DEEP_OVERHEAD
+public static finallong DEEP_OVERHEAD
 
 
 
@@ -2857,7 +2857,7 @@ public staticorg.apache.hadoop.fs.Path
 
 moveCompactedFilesIntoPlace
-privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHStoreFilemoveCompactedFilesIntoPlace(CompactionRequestImplcr,
+privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHStoreFilemoveCompactedFilesIntoPlace(CompactionRequestImplcr,
  https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.PathnewFiles,
  Useruser)
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
@@ -2873,7 +2873,7 @@ public staticorg.apache.hadoop.fs.Path
 
 moveFileIntoPlace
-HStoreFilemoveFileIntoPlace(org.apache.hadoop.fs.PathnewFile)
+HStoreFilemoveFileIntoPlace(org.apache.hadoop.fs.PathnewFile)
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
 Throws:
@@ -2887,7 +2887,7 @@ public staticorg.apache.hadoop.fs.Path
 
 writeCompactionWalRecord
-privatevoidwriteCompactionWalRecord(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionHStoreFilefilesCompacted,
+privatevoidwriteCompactionWalRecord(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionHStoreFilefilesCompacted,
   https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionHStoreFilenewFiles)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Writes the compaction WAL record.
@@ -2906,7 +2906,7 @@ public staticorg.apache.hadoop.fs.Path
 
 replaceStoreFiles
-voidreplaceStoreFiles(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionHStoreFilecompactedFiles,
+voidreplaceStoreFiles(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionHStoreFilecompactedFiles,
https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionHStoreFileresult)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
@@ -2921,7 +2921,7 @@ public staticorg.apache.hadoop.fs.Path
 
 updateSpaceQuotaAfterFileReplacement
-voidupdateSpaceQuotaAfterFileReplacement(RegionSizeStoresizeStore,
+voidupdateSpaceQuotaAfterFileReplacement(RegionSizeStoresizeStore,
   RegionInforegionInfo,
   https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionHStoreFileoldFiles,
   https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionHStoreFilenewFiles)
@@ -2942,7 +2942,7 @@ public staticorg.apache.hadoop.fs.Path
 
 logCompactionEndMessage
-privatevoidlogCompactionEndMessage(CompactionRequestImplcr,
+privatevoidlogCompactionEndMessage(CompactionRequestImplcr,
  https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHStoreFilesfs,
  longnow,
  longcompactionStartTime)
@@ 

[26/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.

2018-08-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/CloseRegionProcedure.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/CloseRegionProcedure.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/CloseRegionProcedure.html
index 8ea4be9..1788f01 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/CloseRegionProcedure.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/CloseRegionProcedure.html
@@ -28,66 +28,72 @@
 020import java.io.IOException;
 021import 
org.apache.hadoop.hbase.ServerName;
 022import 
org.apache.hadoop.hbase.client.RegionInfo;
-023import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
-024import 
org.apache.hadoop.hbase.master.procedure.RSProcedureDispatcher.RegionCloseOperation;
-025import 
org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer;
-026import 
org.apache.hadoop.hbase.procedure2.RemoteProcedureDispatcher.RemoteOperation;
-027import 
org.apache.yetus.audience.InterfaceAudience;
-028import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-029import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CloseRegionProcedureStateData;
-030
-031/**
-032 * The remote procedure used to close a 
region.
-033 */
-034@InterfaceAudience.Private
-035public class CloseRegionProcedure extends 
RegionRemoteProcedureBase {
-036
-037  // For a region move operation, we will 
assign the region after we unassign it, this is the target
-038  // server for the subsequent assign. We 
will send this value to RS, and RS will record the region
-039  // in a Map to tell client that where 
the region has been moved to. Can be null. And also, can be
-040  // wrong(but do not make it wrong 
intentionally). The client can handle this error.
-041  private ServerName assignCandidate;
-042
-043  public CloseRegionProcedure() {
-044super();
-045  }
-046
-047  public CloseRegionProcedure(RegionInfo 
region, ServerName targetServer,
-048  ServerName assignCandidate) {
-049super(region, targetServer);
-050this.assignCandidate = 
assignCandidate;
-051  }
-052
-053  @Override
-054  public TableOperationType 
getTableOperationType() {
-055return 
TableOperationType.REGION_UNASSIGN;
-056  }
-057
-058  @Override
-059  public RemoteOperation 
remoteCallBuild(MasterProcedureEnv env, ServerName remote) {
-060return new RegionCloseOperation(this, 
region, assignCandidate);
-061  }
-062
-063  @Override
-064  protected void 
serializeStateData(ProcedureStateSerializer serializer) throws IOException {
-065
super.serializeStateData(serializer);
-066CloseRegionProcedureStateData.Builder 
builder = CloseRegionProcedureStateData.newBuilder();
-067if (assignCandidate != null) {
-068  
builder.setAssignCandidate(ProtobufUtil.toServerName(assignCandidate));
-069}
-070
serializer.serialize(builder.build());
-071  }
-072
-073  @Override
-074  protected void 
deserializeStateData(ProcedureStateSerializer serializer) throws IOException 
{
-075
super.deserializeStateData(serializer);
-076CloseRegionProcedureStateData data 
=
-077  
serializer.deserialize(CloseRegionProcedureStateData.class);
-078if (data.hasAssignCandidate()) {
-079  assignCandidate = 
ProtobufUtil.toServerName(data.getAssignCandidate());
-080}
-081  }
-082}
+023import 
org.apache.hadoop.hbase.master.RegionState;
+024import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
+025import 
org.apache.hadoop.hbase.master.procedure.RSProcedureDispatcher.RegionCloseOperation;
+026import 
org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer;
+027import 
org.apache.hadoop.hbase.procedure2.RemoteProcedureDispatcher.RemoteOperation;
+028import 
org.apache.yetus.audience.InterfaceAudience;
+029import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
+030import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.CloseRegionProcedureStateData;
+031
+032/**
+033 * The remote procedure used to close a 
region.
+034 */
+035@InterfaceAudience.Private
+036public class CloseRegionProcedure extends 
RegionRemoteProcedureBase {
+037
+038  // For a region move operation, we will 
assign the region after we unassign it, this is the target
+039  // server for the subsequent assign. We 
will send this value to RS, and RS will record the region
+040  // in a Map to tell client that where 
the region has been moved to. Can be null. And also, can be
+041  // wrong(but do not make it wrong 
intentionally). The client can handle this error.
+042  private ServerName assignCandidate;
+043
+044  public CloseRegionProcedure() {
+045super();
+046  }
+047
+048  public CloseRegionProcedure(RegionInfo 
region, ServerName targetServer,
+049  ServerName assignCandidate) {
+050super(region, 

[26/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.

2018-08-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/apidocs/src-html/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html
index 77dc6d6..67b265c 100644
--- 
a/apidocs/src-html/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html
+++ 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html
@@ -29,446 +29,462 @@
 021
 022import java.io.IOException;
 023import java.util.ArrayList;
-024
-025import org.apache.hadoop.hbase.Cell;
-026import 
org.apache.hadoop.hbase.CellUtil;
-027import 
org.apache.hadoop.hbase.CompareOperator;
-028import 
org.apache.hadoop.hbase.PrivateCellUtil;
-029import 
org.apache.yetus.audience.InterfaceAudience;
-030import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-031import 
org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
-032import 
org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException;
-033import 
org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations;
-034import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-035import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-036import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
-037import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.CompareType;
-038import 
org.apache.hadoop.hbase.util.Bytes;
-039
-040import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
-041
-042/**
-043 * This filter is used to filter cells 
based on value. It takes a {@link CompareFilter.CompareOp}
-044 * operator (equal, greater, not equal, 
etc), and either a byte [] value or
-045 * a ByteArrayComparable.
-046 * p
-047 * If we have a byte [] value then we 
just do a lexicographic compare. For
-048 * example, if passed value is 'b' and 
cell has 'a' and the compare operator
-049 * is LESS, then we will filter out this 
cell (return true).  If this is not
-050 * sufficient (eg you want to deserialize 
a long and then compare it to a fixed
-051 * long value), then you can pass in your 
own comparator instead.
-052 * p
-053 * You must also specify a family and 
qualifier.  Only the value of this column
-054 * will be tested. When using this filter 
on a 
-055 * {@link 
org.apache.hadoop.hbase.CellScanner} with specified
-056 * inputs, the column to be tested should 
also be added as input (otherwise
-057 * the filter will regard the column as 
missing).
-058 * p
-059 * To prevent the entire row from being 
emitted if the column is not found
-060 * on a row, use {@link 
#setFilterIfMissing}.
-061 * Otherwise, if the column is found, the 
entire row will be emitted only if
-062 * the value passes.  If the value fails, 
the row will be filtered out.
-063 * p
-064 * In order to test values of previous 
versions (timestamps), set
-065 * {@link #setLatestVersionOnly} to 
false. The default is true, meaning that
-066 * only the latest version's value is 
tested and all previous versions are ignored.
-067 * p
-068 * To filter based on the value of all 
scanned columns, use {@link ValueFilter}.
-069 */
-070@InterfaceAudience.Public
-071public class SingleColumnValueFilter 
extends FilterBase {
-072
-073  protected byte [] columnFamily;
-074  protected byte [] columnQualifier;
-075  protected CompareOperator op;
-076  protected 
org.apache.hadoop.hbase.filter.ByteArrayComparable comparator;
-077  protected boolean foundColumn = 
false;
-078  protected boolean matchedColumn = 
false;
-079  protected boolean filterIfMissing = 
false;
-080  protected boolean latestVersionOnly = 
true;
-081
-082  /**
-083   * Constructor for binary compare of 
the value of a single column.  If the
-084   * column is found and the condition 
passes, all columns of the row will be
-085   * emitted.  If the condition fails, 
the row will not be emitted.
-086   * p
-087   * Use the filterIfColumnMissing flag 
to set whether the rest of the columns
-088   * in a row will be emitted if the 
specified column to check is not found in
-089   * the row.
-090   *
-091   * @param family name of column 
family
-092   * @param qualifier name of column 
qualifier
-093   * @param compareOp operator
-094   * @param value value to compare column 
values against
-095   * @deprecated Since 2.0.0. Will be 
removed in 3.0.0. Use
-096   * {@link 
#SingleColumnValueFilter(byte[], byte[], CompareOperator, byte[])} instead.
-097   */
-098  @Deprecated
-099  public SingleColumnValueFilter(final 
byte [] family, final byte [] qualifier,
-100  final CompareOp compareOp, final 
byte[] value) {
-101this(family, qualifier, 
CompareOperator.valueOf(compareOp.name()),
-102  new 
org.apache.hadoop.hbase.filter.BinaryComparator(value));
-103  }
-104
-105  /**
-106   * Constructor for binary 

[26/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  CollectionHStore stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}
+888
+889MonitoredTask status = 

[26/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.

2018-08-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/assignment/ServerState.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/ServerState.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/ServerState.html
new file mode 100644
index 000..5267fdd
--- /dev/null
+++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/ServerState.html
@@ -0,0 +1,399 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+ServerState (Apache HBase 3.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods = {"i0":9,"i1":9};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Enum Constants|
+Field|
+Method
+
+
+Detail:
+Enum Constants|
+Field|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.master.assignment
+Enum ServerState
+
+
+
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+
+
+https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">java.lang.EnumServerState
+
+
+org.apache.hadoop.hbase.master.assignment.ServerState
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable, https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableServerState
+
+
+
+@InterfaceAudience.Private
+ enum ServerState
+extends https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumServerState
+Server State.
+
+
+
+
+
+
+
+
+
+
+
+Enum Constant Summary
+
+Enum Constants
+
+Enum Constant and Description
+
+
+OFFLINE
+WAL splitting done.
+
+
+
+ONLINE
+Initial state.
+
+
+
+SPLITTING
+Server expired/crashed.
+
+
+
+SPLITTING_META
+Only server which carries meta can have this state.
+
+
+
+SPLITTING_META_DONE
+Indicate that the meta splitting is done.
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsStatic MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+static ServerState
+valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringname)
+Returns the enum constant of this type with the specified 
name.
+
+
+
+static ServerState[]
+values()
+Returns an array containing the constants of this enum 
type, in
+the order they are declared.
+
+
+
+
+
+
+
+Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">Enum
+https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#compareTo-E-;
 title="class or interface in java.lang">compareTo, https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#getDeclaringClass--;
 title="class or interface in java.lang">getDeclaringClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#hashCode--;
 title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/
 8/docs/api/java/lang/Enum.html?is-external=true#name--" title="class or 
interface in java.lang">name, https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#ordinal--;
 title="class or interface in java.lang">ordinal, https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#toString--;
 title="class or interface in java.lang">toString, 

[26/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.

2018-08-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.Importer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.Importer.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.Importer.html
index 39170f0..7859ebc 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.Importer.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.Importer.html
@@ -230,564 +230,567 @@
 222  }
 223}
 224  } catch (InterruptedException e) 
{
-225e.printStackTrace();
-226  }
-227}
-228
-229@Override
-230public void setup(Context context) 
throws IOException {
-231  cfRenameMap = 
createCfRenameMap(context.getConfiguration());
-232  filter = 
instantiateFilter(context.getConfiguration());
-233  int reduceNum = 
context.getNumReduceTasks();
-234  Configuration conf = 
context.getConfiguration();
-235  TableName tableName = 
TableName.valueOf(context.getConfiguration().get(TABLE_NAME));
-236  try (Connection conn = 
ConnectionFactory.createConnection(conf);
-237  RegionLocator regionLocator = 
conn.getRegionLocator(tableName)) {
-238byte[][] startKeys = 
regionLocator.getStartKeys();
-239if (startKeys.length != 
reduceNum) {
-240  throw new IOException("Region 
split after job initialization");
-241}
-242CellWritableComparable[] 
startKeyWraps =
-243new 
CellWritableComparable[startKeys.length - 1];
-244for (int i = 1; i  
startKeys.length; ++i) {
-245  startKeyWraps[i - 1] =
-246  new 
CellWritableComparable(KeyValueUtil.createFirstOnRow(startKeys[i]));
-247}
-248
CellWritableComparablePartitioner.START_KEYS = startKeyWraps;
-249  }
-250}
-251  }
-252
-253  /**
-254   * A mapper that just writes out 
KeyValues.
-255   */
-256  
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="EQ_COMPARETO_USE_OBJECT_EQUALS",
-257  justification="Writables are going 
away and this has been this way forever")
-258  public static class CellImporter 
extends TableMapperImmutableBytesWritable, Cell {
-259private Mapbyte[], byte[] 
cfRenameMap;
-260private Filter filter;
-261private static final Logger LOG = 
LoggerFactory.getLogger(CellImporter.class);
-262
-263/**
-264 * @param row  The current table row 
key.
-265 * @param value  The columns.
-266 * @param context  The current 
context.
-267 * @throws IOException When something 
is broken with the data.
-268 */
-269@Override
-270public void 
map(ImmutableBytesWritable row, Result value,
-271  Context context)
-272throws IOException {
-273  try {
-274if (LOG.isTraceEnabled()) {
-275  LOG.trace("Considering the 
row."
-276  + Bytes.toString(row.get(), 
row.getOffset(), row.getLength()));
-277}
-278if (filter == null
-279|| 
!filter.filterRowKey(PrivateCellUtil.createFirstOnRow(row.get(), 
row.getOffset(),
-280(short) 
row.getLength( {
-281  for (Cell kv : 
value.rawCells()) {
-282kv = filterKv(filter, kv);
-283// skip if we filtered it 
out
-284if (kv == null) continue;
-285context.write(row, new 
MapReduceExtendedCell(convertKv(kv, cfRenameMap)));
-286  }
-287}
-288  } catch (InterruptedException e) 
{
-289e.printStackTrace();
-290  }
-291}
-292
-293@Override
-294public void setup(Context context) 
{
-295  cfRenameMap = 
createCfRenameMap(context.getConfiguration());
-296  filter = 
instantiateFilter(context.getConfiguration());
-297}
-298  }
-299
-300  /**
-301   * Write table content out to files in 
hdfs.
-302   */
-303  public static class Importer extends 
TableMapperImmutableBytesWritable, Mutation {
-304private Mapbyte[], byte[] 
cfRenameMap;
-305private ListUUID 
clusterIds;
-306private Filter filter;
-307private Durability durability;
-308
-309/**
-310 * @param row  The current table row 
key.
-311 * @param value  The columns.
-312 * @param context  The current 
context.
-313 * @throws IOException When something 
is broken with the data.
-314 */
-315@Override
-316public void 
map(ImmutableBytesWritable row, Result value,
-317  Context context)
-318throws IOException {
-319  try {
-320writeResult(row, value, 
context);
-321  } catch (InterruptedException e) 
{
-322e.printStackTrace();
-323  }
-324}
-325
-326private void 
writeResult(ImmutableBytesWritable key, Result result, Context context)
-327throws IOException, 
InterruptedException {
-328  Put put = null;
-329  Delete delete = null;
-330  if (LOG.isTraceEnabled()) {
-331

[26/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.

2018-08-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReaderImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReaderImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReaderImpl.html
index b7b4236..3d1edb3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReaderImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReaderImpl.html
@@ -259,1863 +259,1867 @@
 251   * + Metadata!  + = See note on 
BLOCK_METADATA_SPACE above.
 252   * ++
 253   * /code
-254   * @see #serialize(ByteBuffer)
+254   * @see #serialize(ByteBuffer, 
boolean)
 255   */
-256  static final 
CacheableDeserializerCacheable BLOCK_DESERIALIZER =
-257  new 
CacheableDeserializerCacheable() {
-258@Override
-259public HFileBlock 
deserialize(ByteBuff buf, boolean reuse, MemoryType memType)
-260throws IOException {
-261  // The buf has the file block 
followed by block metadata.
-262  // Set limit to just before the 
BLOCK_METADATA_SPACE then rewind.
-263  buf.limit(buf.limit() - 
BLOCK_METADATA_SPACE).rewind();
-264  // Get a new buffer to pass the 
HFileBlock for it to 'own'.
-265  ByteBuff newByteBuff;
-266  if (reuse) {
-267newByteBuff = buf.slice();
-268  } else {
-269int len = buf.limit();
-270newByteBuff = new 
SingleByteBuff(ByteBuffer.allocate(len));
-271newByteBuff.put(0, buf, 
buf.position(), len);
-272  }
-273  // Read out the 
BLOCK_METADATA_SPACE content and shove into our HFileBlock.
-274  buf.position(buf.limit());
-275  buf.limit(buf.limit() + 
HFileBlock.BLOCK_METADATA_SPACE);
-276  boolean usesChecksum = buf.get() == 
(byte) 1;
-277  long offset = buf.getLong();
-278  int nextBlockOnDiskSize = 
buf.getInt();
-279  HFileBlock hFileBlock =
-280  new HFileBlock(newByteBuff, 
usesChecksum, memType, offset, nextBlockOnDiskSize, null);
-281  return hFileBlock;
-282}
-283
-284@Override
-285public int 
getDeserialiserIdentifier() {
-286  return DESERIALIZER_IDENTIFIER;
-287}
-288
-289@Override
-290public HFileBlock 
deserialize(ByteBuff b) throws IOException {
-291  // Used only in tests
-292  return deserialize(b, false, 
MemoryType.EXCLUSIVE);
-293}
-294  };
-295
-296  private static final int 
DESERIALIZER_IDENTIFIER;
-297  static {
-298DESERIALIZER_IDENTIFIER =
-299
CacheableDeserializerIdManager.registerDeserializer(BLOCK_DESERIALIZER);
-300  }
-301
-302  /**
-303   * Copy constructor. Creates a shallow 
copy of {@code that}'s buffer.
-304   */
-305  private HFileBlock(HFileBlock that) {
-306this(that, false);
-307  }
-308
-309  /**
-310   * Copy constructor. Creates a 
shallow/deep copy of {@code that}'s buffer as per the boolean
-311   * param.
-312   */
-313  private HFileBlock(HFileBlock that, 
boolean bufCopy) {
-314init(that.blockType, 
that.onDiskSizeWithoutHeader,
-315
that.uncompressedSizeWithoutHeader, that.prevBlockOffset,
-316that.offset, 
that.onDiskDataSizeWithHeader, that.nextBlockOnDiskSize, that.fileContext);
-317if (bufCopy) {
-318  this.buf = new 
SingleByteBuff(ByteBuffer.wrap(that.buf.toBytes(0, that.buf.limit(;
-319} else {
-320  this.buf = that.buf.duplicate();
-321}
-322  }
-323
-324  /**
-325   * Creates a new {@link HFile} block 
from the given fields. This constructor
-326   * is used only while writing blocks 
and caching,
-327   * and is sitting in a byte buffer and 
we want to stuff the block into cache.
-328   *
-329   * pTODO: The caller presumes 
no checksumming
-330   * required of this block instance 
since going into cache; checksum already verified on
-331   * underlying block data pulled in from 
filesystem. Is that correct? What if cache is SSD?
+256  public static final 
CacheableDeserializerCacheable BLOCK_DESERIALIZER = new 
BlockDeserializer();
+257
+258  public static final class 
BlockDeserializer implements CacheableDeserializerCacheable {
+259private BlockDeserializer() {
+260}
+261
+262@Override
+263public HFileBlock 
deserialize(ByteBuff buf, boolean reuse, MemoryType memType)
+264throws IOException {
+265  // The buf has the file block 
followed by block metadata.
+266  // Set limit to just before the 
BLOCK_METADATA_SPACE then rewind.
+267  buf.limit(buf.limit() - 
BLOCK_METADATA_SPACE).rewind();
+268  // Get a new buffer to pass the 
HFileBlock for it to 'own'.
+269  ByteBuff newByteBuff;
+270  if (reuse) {
+271newByteBuff = buf.slice();
+272  } else {
+273int len = buf.limit();
+274newByteBuff = new 
SingleByteBuff(ByteBuffer.allocate(len));
+275newByteBuff.put(0, buf, 
buf.position(), len);

[26/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.

2018-07-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
index 0de3ee4..1cd2bd5 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
@@ -157,6 +157,11 @@
 MasterCoprocessorHost.preAbortProcedure(ProcedureExecutorMasterProcedureEnvprocEnv,
  longprocId)
 
+
+private long
+ClusterSchemaServiceImpl.submitProcedure(ProcedureMasterProcedureEnvprocedure,
+   NonceKeynonceKey)
+
 
 
 
@@ -202,17 +207,13 @@
 
 
 protected Procedure.LockState
-GCRegionProcedure.acquireLock(MasterProcedureEnvenv)
-
-
-protected Procedure.LockState
 MergeTableRegionsProcedure.acquireLock(MasterProcedureEnvenv)
 
-
+
 protected Procedure.LockState
 RegionTransitionProcedure.acquireLock(MasterProcedureEnvenv)
 
-
+
 protected boolean
 RegionTransitionProcedure.addToRemoteDispatcher(MasterProcedureEnvenv,
  ServerNametargetServer)
@@ -222,7 +223,7 @@
  to pick it up.
 
 
-
+
 private void
 SplitTableRegionProcedure.checkClosedRegion(MasterProcedureEnvenv)
 Check whether there is recovered.edits in the closed region
@@ -230,7 +231,7 @@
  to abort region split to prevent data loss
 
 
-
+
 private void
 MergeTableRegionsProcedure.checkClosedRegion(MasterProcedureEnvenv,
  RegionInforegionInfo)
@@ -239,19 +240,19 @@
  to abort region merge to prevent data loss
 
 
-
+
 private void
 MergeTableRegionsProcedure.checkClosedRegions(MasterProcedureEnvenv)
 check the closed regions
 
 
-
+
 private static void
 MergeTableRegionsProcedure.checkRegionsToMerge(MasterProcedureEnvenv,
RegionInfo[]regionsToMerge,
booleanforcible)
 
-
+
 private static void
 MergeTableRegionsProcedure.checkRegionsToMerge(MasterProcedureEnvenv,
RegionInforegionToMergeA,
@@ -260,7 +261,7 @@
 One time checks.
 
 
-
+
 private void
 SplitTableRegionProcedure.checkSplittable(MasterProcedureEnvenv,
RegionInforegionToSplit,
@@ -268,113 +269,113 @@
 Check whether the region is splittable
 
 
-
+
 private void
 MergeTableRegionsProcedure.cleanupMergedRegion(MasterProcedureEnvenv)
 Clean up a merged region
 
 
-
+
 private AssignProcedure[]
 SplitTableRegionProcedure.createAssignProcedures(MasterProcedureEnvenv,
   intregionReplication)
 
-
+
 private AssignProcedure[]
 MergeTableRegionsProcedure.createAssignProcedures(MasterProcedureEnvenv,
   intregionReplication)
 
-
+
 void
 SplitTableRegionProcedure.createDaughterRegions(MasterProcedureEnvenv)
 Create daughter regions
 
 
-
+
 private GCRegionProcedure[]
 GCMergedRegionsProcedure.createGCRegionProcedures(MasterProcedureEnvenv)
 
-
+
 private void
 MergeTableRegionsProcedure.createMergedRegion(MasterProcedureEnvenv)
 Create a merged region
 
 
-
+
 private UnassignProcedure[]
 SplitTableRegionProcedure.createUnassignProcedures(MasterProcedureEnvenv,
 intregionReplication)
 
-
+
 private UnassignProcedure[]
 MergeTableRegionsProcedure.createUnassignProcedures(MasterProcedureEnvenv,
 intregionReplication)
 
-
+
 protected Procedure[]
 RegionTransitionProcedure.execute(MasterProcedureEnvenv)
 
-
+
 protected StateMachineProcedure.Flow
 GCMergedRegionsProcedure.executeFromState(MasterProcedureEnvenv,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.GCMergedRegionsStatestate)
 
-
+
 protected StateMachineProcedure.Flow
 GCRegionProcedure.executeFromState(MasterProcedureEnvenv,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.GCRegionStatestate)
 
-
+
 protected StateMachineProcedure.Flow
 MergeTableRegionsProcedure.executeFromState(MasterProcedureEnvenv,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsStatestate)
 
-
+
 protected StateMachineProcedure.Flow
 MoveRegionProcedure.executeFromState(MasterProcedureEnvenv,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MoveRegionStatestate)
 
-
+
 protected StateMachineProcedure.Flow
 SplitTableRegionProcedure.executeFromState(MasterProcedureEnvenv,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionStatestate)
 
-
+
 protected void
 UnassignProcedure.finishTransition(MasterProcedureEnvenv,
 RegionStates.RegionStateNoderegionNode)
 
-
+
 protected void
 AssignProcedure.finishTransition(MasterProcedureEnvenv,
 

[26/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.

2018-07-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadFunction.html
index 233dba3..91b9055 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadFunction.html
@@ -540,1205 +540,1204 @@
 532  
sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) - {
 533DequeBalancerRegionLoad 
rLoads = oldLoads.get(Bytes.toString(regionName));
 534if (rLoads == null) {
-535  // There was nothing there
-536  rLoads = new 
ArrayDeque();
-537} else if (rLoads.size() = 
numRegionLoadsToRemember) {
-538  rLoads.remove();
-539}
-540rLoads.add(new 
BalancerRegionLoad(rm));
-541
loads.put(Bytes.toString(regionName), rLoads);
-542  });
-543});
-544
-545for(CostFromRegionLoadFunction cost : 
regionLoadFunctions) {
-546  cost.setLoads(loads);
-547}
-548  }
-549
-550  protected void initCosts(Cluster 
cluster) {
-551for (CostFunction c:costFunctions) 
{
-552  c.init(cluster);
-553}
-554  }
-555
-556  protected void 
updateCostsWithAction(Cluster cluster, Action action) {
-557for (CostFunction c : costFunctions) 
{
-558  c.postAction(action);
-559}
-560  }
-561
-562  /**
-563   * Get the names of the cost 
functions
-564   */
-565  public String[] getCostFunctionNames() 
{
-566if (costFunctions == null) return 
null;
-567String[] ret = new 
String[costFunctions.length];
-568for (int i = 0; i  
costFunctions.length; i++) {
-569  CostFunction c = 
costFunctions[i];
-570  ret[i] = 
c.getClass().getSimpleName();
-571}
-572
-573return ret;
-574  }
-575
-576  /**
-577   * This is the main cost function.  It 
will compute a cost associated with a proposed cluster
-578   * state.  All different costs will be 
combined with their multipliers to produce a double cost.
-579   *
-580   * @param cluster The state of the 
cluster
-581   * @param previousCost the previous 
cost. This is used as an early out.
-582   * @return a double of a cost 
associated with the proposed cluster state.  This cost is an
-583   * aggregate of all individual 
cost functions.
-584   */
-585  protected double computeCost(Cluster 
cluster, double previousCost) {
-586double total = 0;
-587
-588for (int i = 0; i  
costFunctions.length; i++) {
-589  CostFunction c = 
costFunctions[i];
-590  this.tempFunctionCosts[i] = 0.0;
-591
-592  if (c.getMultiplier() = 0) {
-593continue;
-594  }
-595
-596  Float multiplier = 
c.getMultiplier();
-597  Double cost = c.cost();
-598
-599  this.tempFunctionCosts[i] = 
multiplier*cost;
-600  total += 
this.tempFunctionCosts[i];
-601
-602  if (total  previousCost) {
-603break;
-604  }
-605}
-606
-607return total;
-608  }
-609
-610  /** Generates a candidate action to be 
applied to the cluster for cost function search */
-611  abstract static class 
CandidateGenerator {
-612abstract Cluster.Action 
generate(Cluster cluster);
-613
-614/**
-615 * From a list of regions pick a 
random one. Null can be returned which
-616 * {@link 
StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region 
move
-617 * rather than swap.
-618 *
-619 * @param clusterThe state of 
the cluster
-620 * @param server index of the 
server
-621 * @param chanceOfNoSwap Chance that 
this will decide to try a move rather
-622 *   than a 
swap.
-623 * @return a random {@link 
RegionInfo} or null if an asymmetrical move is
-624 * suggested.
-625 */
-626protected int 
pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) {
-627  // Check to see if this is just a 
move.
-628  if 
(cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat()  
chanceOfNoSwap) {
-629// signal a move only.
-630return -1;
-631  }
-632  int rand = 
RANDOM.nextInt(cluster.regionsPerServer[server].length);
-633  return 
cluster.regionsPerServer[server][rand];
-634
-635}
-636protected int 
pickRandomServer(Cluster cluster) {
-637  if (cluster.numServers  1) {
-638return -1;
-639  }
-640
-641  return 
RANDOM.nextInt(cluster.numServers);
-642}
-643
-644protected int pickRandomRack(Cluster 
cluster) {
-645  if (cluster.numRacks  1) {
-646return 

[26/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.

2018-07-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/MasterNotRunningException.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/MasterNotRunningException.html 
b/apidocs/org/apache/hadoop/hbase/MasterNotRunningException.html
index f6b49a7..23a0240 100644
--- a/apidocs/org/apache/hadoop/hbase/MasterNotRunningException.html
+++ b/apidocs/org/apache/hadoop/hbase/MasterNotRunningException.html
@@ -1,6 +1,6 @@
 http://www.w3.org/TR/html4/loose.dtd;>
 
-
+
 
 
 
@@ -20,38 +20,38 @@
 //-->
 
 
-您的浏览器已禁用 JavaScript。
+JavaScript is disabled on your browser.
 
 
 
 
 
-跳过导航链接
+Skip navigation links
 
 
 
-
-概览
-程序包
-ç±»
-使用
-树
-已过时
-索引
-帮助
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
 
 
 
 
-上一个类
-下一个类
+PrevClass
+NextClass
 
 
-框架
-无框架
+Frames
+NoFrames
 
 
-所有类
+AllClasses
 
 
 
 
org.apache.hadoop.hbase
-

ç±» MasterNotRunningException

+

Class MasterNotRunningException

    -
  • https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="java.lang中的类或接口">java.lang.Object
  • +
  • https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true"; title="class or interface in java.lang">java.lang.Object
    • -
    • https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true"; title="java.lang中的类或接口">java.lang.Throwable
    • +
    • https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true"; title="class or interface in java.lang">java.lang.Throwable
      • -
      • https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true"; title="java.lang中的类或接口">java.lang.Exception
      • +
      • https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true"; title="class or interface in java.lang">java.lang.Exception
        • -
        • https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true"; title="java.io中的类或接口">java.io.IOException
        • +
        • https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true"; title="class or interface in java.io">java.io.IOException
          • -
          • org.apache.hadoop.hbase.HBaseIOException
          • +
          • [26/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
            --
            diff --git a/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html 
            b/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
            index 8a6ff31..f1708c1 100644
            --- a/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
            +++ b/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
            @@ -1,6 +1,6 @@
             http://www.w3.org/TR/html4/loose.dtd;>
             
            -
            +
             
             
             
            @@ -19,45 +19,45 @@
             }
             //-->
             var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":9,"i15":10,"i16":9,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10};
            -var tabs = {65535:["t0","All Methods"],1:["t1","Static 
            Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
            +var tabs = 
            {65535:["t0","所有方法"],1:["t1","静态方法"],2:["t2","实例方法"],8:["t4","å
            …·ä½“方法"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
             var tableTab = "tableTab";
             var activeTableTab = "activeTableTab";
             
             
            -JavaScript is disabled on your browser.
            +您的浏览器已禁用 JavaScript。
             
             
             
             
             
            -Skip navigation links
            +跳过导航链接
             
             
             
            -
            -Overview
            -Package
            -Class
            -Use
            -Tree
            -Deprecated
            -Index
            -Help
            +
            +概览
            +程序包
            +ç±»
            +使用
            +树
            +已过时
            +索引
            +帮助
             
             
             
             
            -PrevClass
            -NextClass
            +上一个类
            +下一个类
             
             
            -Frames
            -NoFrames
            +框架
            +无框架
             
             
            -AllClasses
            +所有类
             
             
             

            [26/51] [partial] hbase-site git commit: Published site at 0f23784182ab88649de340d75804e0ff20dcd0fc.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/ThreadSafeMemStoreSizing.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/ThreadSafeMemStoreSizing.html 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/ThreadSafeMemStoreSizing.html
            index b497649..fc09302 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/ThreadSafeMemStoreSizing.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/ThreadSafeMemStoreSizing.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10};
            +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10};
             var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -202,28 +202,33 @@ implements Method and Description
             
             
            +boolean
            +compareAndSetDataSize(longexpected,
            + longupdated)
            +
            +
             long
             getDataSize()
             
            -
            +
             long
             getHeapSize()
             
            -
            +
             MemStoreSize
             getMemStoreSize()
             
            -
            +
             long
             getOffHeapSize()
             
            -
            +
             long
             incMemStoreSize(longdataSizeDelta,
            longheapSizeDelta,
            longoffHeapSizeDelta)
             
            -
            +
             https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String
             toString()
             
            @@ -361,13 +366,27 @@ implements 
            +
            +
            +
            +
            +compareAndSetDataSize
            +publicbooleancompareAndSetDataSize(longexpected,
            + longupdated)
            +
            +Specified by:
            +compareAndSetDataSizein
             interfaceMemStoreSizing
            +
            +
            +
             
             
             
             
             
             getDataSize
            -publiclonggetDataSize()
            +publiclonggetDataSize()
             
             Specified by:
             getDataSizein
             interfaceMemStoreSizing
            @@ -380,7 +399,7 @@ implements 
             
             getHeapSize
            -publiclonggetHeapSize()
            +publiclonggetHeapSize()
             
             Specified by:
             getHeapSizein
             interfaceMemStoreSizing
            @@ -393,7 +412,7 @@ implements 
             
             getOffHeapSize
            -publiclonggetOffHeapSize()
            +publiclonggetOffHeapSize()
             
             Specified by:
             getOffHeapSizein
             interfaceMemStoreSizing
            @@ -406,7 +425,7 @@ implements 
             
             toString
            -publichttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringtoString()
            +publichttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringtoString()
             
             Overrides:
             https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
             title="class or interface in java.lang">toStringin 
            classhttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CSLMImmutableSegment.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CSLMImmutableSegment.html
             
            b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CSLMImmutableSegment.html
            index e4058a7..dd90956 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CSLMImmutableSegment.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CSLMImmutableSegment.html
            @@ -119,7 +119,7 @@
             
             
             CellArrayImmutableSegment(CSLMImmutableSegmentsegment,
            - MemStoreSizingmemstoreSizing,
            + MemStoreSizingmss,
              MemStoreCompactionStrategy.Actionaction)
             
              C-tor to be used when new CellChunkImmutableSegment is built as a result of 
            flattening
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CellSet.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CellSet.html 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CellSet.html
            index 531bfad..9654ae2 100644
            --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CellSet.html
            +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CellSet.html
            @@ -141,9 +141,10 @@
             
             
             private void
            -CellChunkImmutableSegment.reinitializeCellSet(intnumOfCells,
            +CellChunkImmutableSegment.reinitializeCellSet(intnumOfCells,
            KeyValueScannersegmentScanner,
            CellSetoldCellSet,
            +   MemStoreSizingmemstoreSizing,
            MemStoreCompactionStrategy.Actionaction)
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/CompactingMemStore.InMemoryCompactionRunnable.html
            --
            diff --git 
            

            [26/51] [partial] hbase-site git commit: Published site at 85b41f36e01214b6485c9352875c84ebf877dab3.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
             
            b/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
            index 5731302..08a3446 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
            @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private static enum AccessController.OpType
            +private static enum AccessController.OpType
             extends https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
             title="class or interface in java.lang">EnumAccessController.OpType
             
             
            @@ -257,7 +257,7 @@ the order they are declared.
             
             
             GET
            -public static finalAccessController.OpType GET
            +public static finalAccessController.OpType GET
             
             
             
            @@ -266,7 +266,7 @@ the order they are declared.
             
             
             EXISTS
            -public static finalAccessController.OpType EXISTS
            +public static finalAccessController.OpType EXISTS
             
             
             
            @@ -275,7 +275,7 @@ the order they are declared.
             
             
             SCAN
            -public static finalAccessController.OpType SCAN
            +public static finalAccessController.OpType SCAN
             
             
             
            @@ -284,7 +284,7 @@ the order they are declared.
             
             
             PUT
            -public static finalAccessController.OpType PUT
            +public static finalAccessController.OpType PUT
             
             
             
            @@ -293,7 +293,7 @@ the order they are declared.
             
             
             DELETE
            -public static finalAccessController.OpType DELETE
            +public static finalAccessController.OpType DELETE
             
             
             
            @@ -302,7 +302,7 @@ the order they are declared.
             
             
             CHECK_AND_PUT
            -public static finalAccessController.OpType CHECK_AND_PUT
            +public static finalAccessController.OpType CHECK_AND_PUT
             
             
             
            @@ -311,7 +311,7 @@ the order they are declared.
             
             
             CHECK_AND_DELETE
            -public static finalAccessController.OpType CHECK_AND_DELETE
            +public static finalAccessController.OpType CHECK_AND_DELETE
             
             
             
            @@ -320,7 +320,7 @@ the order they are declared.
             
             
             INCREMENT_COLUMN_VALUE
            -public static finalAccessController.OpType INCREMENT_COLUMN_VALUE
            +public static finalAccessController.OpType INCREMENT_COLUMN_VALUE
             
             
             
            @@ -329,7 +329,7 @@ the order they are declared.
             
             
             APPEND
            -public static finalAccessController.OpType APPEND
            +public static finalAccessController.OpType APPEND
             
             
             
            @@ -338,7 +338,7 @@ the order they are declared.
             
             
             INCREMENT
            -public static finalAccessController.OpType INCREMENT
            +public static finalAccessController.OpType INCREMENT
             
             
             
            @@ -355,7 +355,7 @@ the order they are declared.
             
             
             type
            -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String type
            +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String type
             
             
             
            @@ -372,7 +372,7 @@ the order they are declared.
             
             
             values
            -public staticAccessController.OpType[]values()
            +public staticAccessController.OpType[]values()
             Returns an array containing the constants of this enum 
            type, in
             the order they are declared.  This method may be used to iterate
             over the constants as follows:
            @@ -392,7 +392,7 @@ for (AccessController.OpType c : 
            AccessController.OpType.values())
             
             
             valueOf
            -public staticAccessController.OpTypevalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringname)
            +public staticAccessController.OpTypevalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringname)
             Returns the enum constant of this type with the specified 
            name.
             The string must match exactly an identifier used to declare an
             enum constant in this type.  (Extraneous whitespace characters are 
            @@ -414,7 +414,7 @@ not permitted.)
             
             
             toString
            -publichttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringtoString()
            +publichttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringtoString()
             
             Overrides:
             https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#toString--;
             title="class or interface in java.lang">toStringin 
            classhttps://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
             title="class or interface in java.lang">EnumAccessController.OpType
            
            
            

            [26/51] [partial] hbase-site git commit: Published site at 6198e1fc7dfa85c3bc6b2855f9a5fb5f4b2354ff.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html
            index 5d06422..d1a589e 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":9,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":42,"i29":42,"i30":42,"i31":42,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":42,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":9,"i84":10,"i85":10,"i86":10,"i87":10,"i88":9,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":41,"i99":10,"i100":10,"i101":10,"i102":10,"i103":42,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109
             
            ":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":9,"i119":10,"i120":10,"i121":10,"i122":42,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":42,"i175":10,"i176":10,"i177":10,"i178":10,"i179":10,"i180":10,"i181":10,"i182":10,"i183":10,"i184":10,"i185":10,"i186":10,"i187":10,"i188":10,"i189":10,"i190":10,"i191":10,"i192":10,"i193":10,"i194":10,"i195":10,"i196":10,"i197":10,"i198":10,"i199":10,"i200":10,"i201":42,"i202":10,"i203":10,"i204":10,"i205":10,"i206":10,"i207":10,"i208":10,"i2
             
            09":10,"i210":10,"i211":10,"i212":10,"i213":10,"i214":10,"i215":10,"i216":10,"i217":10,"i218":10,"i219":10,"i220":10,"i221":10,"i222":10,"i223":10,"i224":10,"i225":10,"i226":10,"i227":10,"i228":10,"i229":10,"i230":10,"i231":10,"i232":10,"i233":10};
            +var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":9,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":42,"i29":42,"i30":42,"i31":42,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":42,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":9,"i84":10,"i85":10,"i86":10,"i87":10,"i88":9,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":41,"i99":10,"i100":10,"i101":10,"i102":10,"i103":42,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109
             
            ":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":9,"i119":10,"i120":10,"i121":10,"i122":42,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":42,"i175":10,"i176":10,"i177":10,"i178":10,"i179":10,"i180":10,"i181":10,"i182":10,"i183":10,"i184":10,"i185":10,"i186":10,"i187":10,"i188":10,"i189":10,"i190":10,"i191":10,"i192":10,"i193":10,"i194":10,"i195":10,"i196":10,"i197":10,"i198":10,"i199":10,"i200":10,"i201":42,"i202":10,"i203":10,"i204":10,"i205":10,"i206":10,"i207":10,"i208":10,"i2
             
            

            [26/51] [partial] hbase-site git commit: Published site at 14087cc919da9f2e0b1a68f701f6365ad9d1d71f.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
             
            b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
            index e574b69..5c20e64 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
            @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -abstract static class StochasticLoadBalancer.CandidateGenerator
            +abstract static class StochasticLoadBalancer.CandidateGenerator
             extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
             Generates a candidate action to be applied to the cluster 
            for cost function search
             
            @@ -232,7 +232,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             CandidateGenerator
            -CandidateGenerator()
            +CandidateGenerator()
             
             
             
            @@ -249,7 +249,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             generate
            -abstractBaseLoadBalancer.Cluster.Actiongenerate(BaseLoadBalancer.Clustercluster)
            +abstractBaseLoadBalancer.Cluster.Actiongenerate(BaseLoadBalancer.Clustercluster)
             
             
             
            @@ -258,7 +258,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             pickRandomRegion
            -protectedintpickRandomRegion(BaseLoadBalancer.Clustercluster,
            +protectedintpickRandomRegion(BaseLoadBalancer.Clustercluster,
            intserver,
            doublechanceOfNoSwap)
             From a list of regions pick a random one. Null can be 
            returned which
            @@ -282,7 +282,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             pickRandomServer
            -protectedintpickRandomServer(BaseLoadBalancer.Clustercluster)
            +protectedintpickRandomServer(BaseLoadBalancer.Clustercluster)
             
             
             
            @@ -291,7 +291,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             pickRandomRack
            -protectedintpickRandomRack(BaseLoadBalancer.Clustercluster)
            +protectedintpickRandomRack(BaseLoadBalancer.Clustercluster)
             
             
             
            @@ -300,7 +300,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             pickOtherRandomServer
            -protectedintpickOtherRandomServer(BaseLoadBalancer.Clustercluster,
            +protectedintpickOtherRandomServer(BaseLoadBalancer.Clustercluster,
             intserverIndex)
             
             
            @@ -310,7 +310,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             pickOtherRandomRack
            -protectedintpickOtherRandomRack(BaseLoadBalancer.Clustercluster,
            +protectedintpickOtherRandomRack(BaseLoadBalancer.Clustercluster,
               intrackIndex)
             
             
            @@ -320,7 +320,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             pickRandomRegions
            -protectedBaseLoadBalancer.Cluster.ActionpickRandomRegions(BaseLoadBalancer.Clustercluster,
            +protectedBaseLoadBalancer.Cluster.ActionpickRandomRegions(BaseLoadBalancer.Clustercluster,
             
            intthisServer,
             
            intotherServer)
             
            @@ -331,7 +331,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             getAction
            -protectedBaseLoadBalancer.Cluster.ActiongetAction(intfromServer,
            +protectedBaseLoadBalancer.Cluster.ActiongetAction(intfromServer,
             intfromRegion,
             inttoServer,
             inttoRegion)
            @@ -343,7 +343,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             getRandomIterationOrder
            -protectedhttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
             title="class or interface in java.lang">IntegergetRandomIterationOrder(intlength)
            +protectedhttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
             title="class or interface in java.lang">IntegergetRandomIterationOrder(intlength)
             Returns a random iteration order of indexes of an array 
            with size length
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
            

            [26/51] [partial] hbase-site git commit: Published site at 72784c2d836a4b977667449d3adec5e8d15453f5.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.EnableTableFuture.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.EnableTableFuture.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.EnableTableFuture.html
            index b6e7636..592c2cc 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.EnableTableFuture.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.EnableTableFuture.html
            @@ -356,3901 +356,3924 @@
             348  public FutureVoid 
            modifyTableAsync(TableDescriptor td) throws IOException {
             349ModifyTableResponse response = 
            executeCallable(
             350  new 
            MasterCallableModifyTableResponse(getConnection(), 
            getRpcControllerFactory()) {
            -351@Override
            -352protected ModifyTableResponse 
            rpcCall() throws Exception {
            -353  
            setPriority(td.getTableName());
            -354  ModifyTableRequest request = 
            RequestConverter.buildModifyTableRequest(
            -355td.getTableName(), td, 
            ng.getNonceGroup(), ng.newNonce());
            -356  return 
            master.modifyTable(getRpcController(), request);
            -357}
            -358  });
            -359return new ModifyTableFuture(this, 
            td.getTableName(), response);
            -360  }
            -361
            -362  @Override
            -363  public ListTableDescriptor 
            listTableDescriptorsByNamespace(byte[] name) throws IOException {
            -364return executeCallable(new 
            MasterCallableListTableDescriptor(getConnection(),
            -365getRpcControllerFactory()) {
            -366  @Override
            -367  protected 
            ListTableDescriptor rpcCall() throws Exception {
            -368return 
            master.listTableDescriptorsByNamespace(getRpcController(),
            -369
            ListTableDescriptorsByNamespaceRequest.newBuilder()
            -370  
            .setNamespaceName(Bytes.toString(name)).build())
            -371.getTableSchemaList()
            -372.stream()
            -373
            .map(ProtobufUtil::toTableDescriptor)
            -374
            .collect(Collectors.toList());
            -375  }
            -376});
            -377  }
            -378
            -379  @Override
            -380  public ListTableDescriptor 
            listTableDescriptors(ListTableName tableNames) throws IOException {
            -381return executeCallable(new 
            MasterCallableListTableDescriptor(getConnection(),
            -382getRpcControllerFactory()) {
            -383  @Override
            -384  protected 
            ListTableDescriptor rpcCall() throws Exception {
            -385GetTableDescriptorsRequest req 
            =
            -386
            RequestConverter.buildGetTableDescriptorsRequest(tableNames);
            -387  return 
            ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(),
            -388  req));
            -389  }
            -390});
            -391  }
            -392
            -393  @Override
            -394  public ListRegionInfo 
            getRegions(final ServerName sn) throws IOException {
            -395AdminService.BlockingInterface admin 
            = this.connection.getAdmin(sn);
            -396// TODO: There is no timeout on this 
            controller. Set one!
            -397HBaseRpcController controller = 
            rpcControllerFactory.newController();
            -398return 
            ProtobufUtil.getOnlineRegions(controller, admin);
            -399  }
            -400
            -401  @Override
            -402  public ListRegionInfo 
            getRegions(TableName tableName) throws IOException {
            -403if 
            (TableName.isMetaTableName(tableName)) {
            -404  return 
            Arrays.asList(RegionInfoBuilder.FIRST_META_REGIONINFO);
            -405} else {
            -406  return 
            MetaTableAccessor.getTableRegions(connection, tableName, true);
            -407}
            -408  }
            -409
            -410  private static class 
            AbortProcedureFuture extends ProcedureFutureBoolean {
            -411private boolean isAbortInProgress;
            -412
            -413public AbortProcedureFuture(
            -414final HBaseAdmin admin,
            -415final Long procId,
            -416final Boolean abortProcResponse) 
            {
            -417  super(admin, procId);
            -418  this.isAbortInProgress = 
            abortProcResponse;
            -419}
            -420
            -421@Override
            -422public Boolean get(long timeout, 
            TimeUnit unit)
            -423throws InterruptedException, 
            ExecutionException, TimeoutException {
            -424  if (!this.isAbortInProgress) {
            -425return false;
            -426  }
            -427  super.get(timeout, unit);
            -428  return true;
            -429}
            -430  }
            -431
            -432  /** @return Connection used by this 
            object. */
            -433  @Override
            -434  public Connection getConnection() {
            -435return connection;
            -436  }
            -437
            -438  @Override
            -439  public boolean tableExists(final 
            TableName tableName) throws IOException {
            -440return executeCallable(new 
            RpcRetryingCallableBoolean() {
            -441  @Override
            -442  protected Boolean rpcCall(int 
            callTimeout) throws Exception {
            -443return 
            MetaTableAccessor.tableExists(connection, tableName);
            -444  }
            -445});
            -446  }
            -447
            -448  @Override
            -449  public HTableDescriptor[] listTables() 
            throws IOException {
            -450return listTables((Pattern)null, 
            false);
            -451  }
            -452
            -453  @Override
            -454  public HTableDescriptor[] 
            listTables(Pattern pattern) throws IOException {
            -455   

            [26/51] [partial] hbase-site git commit: Published site at 9101fc246f86445006bfbcdfda5cc495016dc280.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html 
            b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
            index 61707dd..4a4d446 100644
            --- a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
            +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
            @@ -257,21 +257,21 @@
             
             
             boolean
            -AssignmentManager.waitMetaInitialized(Procedureproc)
            +AssignmentManager.waitMetaAssigned(Procedure?proc,
            +RegionInforegionInfo)
            +Notice that this event does not mean the AM has already 
            finished region state rebuilding.
            +
             
             
             boolean
            -AssignmentManager.waitMetaInitialized(Procedureproc,
            -   RegionInforegionInfo)
            +AssignmentManager.waitMetaLoaded(Procedure?proc)
            +Wait until AM finishes the meta loading, i.e, the region 
            states rebuilding.
            +
             
             
            -boolean
            -AssignmentManager.waitMetaLoaded(Procedureproc)
            -
            -
             protected boolean
             AssignmentManager.waitServerReportEvent(ServerNameserverName,
            - Procedureproc)
            + Procedure?proc)
             
             
             
            @@ -373,38 +373,47 @@
             
             
             class
            +InitMetaProcedure
            +This procedure is used to initialize meta table for a new 
            hbase deploy.
            +
            +
            +
            +class
             ModifyNamespaceProcedure
             The procedure to add a namespace to an existing table.
             
             
            -
            +
             class
             ModifyTableProcedure
             
            -
            +
             class
             RecoverMetaProcedure
            -This procedure recovers meta from prior shutdown/ crash of 
            a server, and brings meta online by
            - assigning meta region/s.
            +Deprecated.
            +Do not use any more, leave 
            it here only for compatible. The recovery work will be
            + done in ServerCrashProcedure 
            directly, and the initial work for meta table
            + will be done by InitMetaProcedure.
            +
             
             
            -
            +
             class
             ReopenTableRegionsProcedure
             Used for non table procedures to reopen the regions for a 
            table.
             
             
            -
            +
             class
             RestoreSnapshotProcedure
             
            -
            +
             class
             ServerCrashProcedure
             Handle crashed server.
             
             
            -
            +
             class
             TruncateTableProcedure
             
            @@ -542,7 +551,9 @@
             
             
             (package private) boolean
            -MetaQueue.requireExclusiveLock(Procedure?proc)
            +MetaQueue.requireExclusiveLock(Procedure?proc)
            +Deprecated.
            +
             
             
             boolean
            @@ -580,64 +591,59 @@
            Procedure?procedure)
             
             
            -boolean
            -MasterProcedureEnv.waitFailoverCleanup(Procedure?proc)
            -
            -
             static byte[]
             ProcedureSyncWait.waitForProcedureToComplete(ProcedureExecutorMasterProcedureEnvprocExec,
               Procedure?proc,
               longtimeout)
             
            -
            +
             static byte[]
             ProcedureSyncWait.waitForProcedureToCompleteIOE(ProcedureExecutorMasterProcedureEnvprocExec,
              Procedure?proc,
              longtimeout)
             
            -
            +
             boolean
             MasterProcedureEnv.waitInitialized(Procedure?proc)
             
            -
            +
             boolean
             MasterProcedureScheduler.waitMetaExclusiveLock(Procedure?procedure)
            -Try to acquire the exclusive lock on meta.
            +Deprecated.
            +only used for RecoverMetaProcedure.
             Should be removed along with
            + RecoverMetaProcedure.
            +
             
             
            -
            +
             boolean
            -MasterProcedureScheduler.waitNamespaceExclusiveLock(Procedureprocedure,
            +MasterProcedureScheduler.waitNamespaceExclusiveLock(Procedure?procedure,
               https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringnamespace)
             Suspend the procedure if the specified namespace is already 
            locked.
             
             
            -
            +
             boolean
             MasterProcedureScheduler.waitPeerExclusiveLock(Procedure?procedure,
              https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringpeerId)
             Try to acquire the exclusive lock on the specified 
            peer.
             
             
            -
            +
             boolean
            -MasterProcedureScheduler.waitRegion(Procedureprocedure,
            +MasterProcedureScheduler.waitRegion(Procedure?procedure,
               RegionInforegionInfo)
             Suspend the procedure if the specified region is already 
            locked.
             
             
            -
            +
             boolean
            -MasterProcedureScheduler.waitRegions(Procedureprocedure,
            +MasterProcedureScheduler.waitRegions(Procedure?procedure,
            TableNametable,
            RegionInfo...regionInfo)
             Suspend the procedure if the specified set of regions are 
            already locked.
             
             
            -
            -boolean
            -MasterProcedureEnv.waitServerCrashProcessingEnabled(Procedure?proc)
            -
             
             boolean
             MasterProcedureScheduler.waitServerExclusiveLock(Procedure?procedure,
            @@ -647,7 +653,7 @@
             
             
             boolean
            -MasterProcedureScheduler.waitTableExclusiveLock(Procedureprocedure,
            +MasterProcedureScheduler.waitTableExclusiveLock(Procedure?procedure,
               TableNametable)
             Suspend the procedure if the specified table is already 
            locked.
             
            @@ -659,7 +665,7 @@
             
             
             boolean
            -MasterProcedureScheduler.waitTableSharedLock(Procedureprocedure,
            

            [26/51] [partial] hbase-site git commit: Published site at 0b28155d274910b4e667b949d51f78809a1eff0b.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.RepairCommand.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.RepairCommand.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.RepairCommand.html
            index f236300..513d2ad 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.RepairCommand.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.RepairCommand.html
            @@ -26,1048 +26,1115 @@
             018
             019package 
            org.apache.hadoop.hbase.backup.impl;
             020
            -021import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
            -022import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
            -023import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
            -024import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
            -025import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
            -026import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
            -027import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
            -028import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
            -029import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
            -030import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
            -031import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
            -032import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
            -033import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
            -034import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
            -035import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
            -036import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
            -037import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
            -038import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
            -039
            -040import java.io.IOException;
            -041import java.net.URI;
            -042import java.util.List;
            +021import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BACKUP_LIST_DESC;
            +022import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
            +023import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
            +024import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
            +025import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
            +026import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP;
            +027import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP_DESC;
            +028import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_LIST;
            +029import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
            +030import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
            +031import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
            +032import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
            +033import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
            +034import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
            +035import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
            +036import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
            +037import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
            +038import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
            +039import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
            +040import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
            +041import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
            +042import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
             043
            -044import 
            org.apache.commons.lang3.StringUtils;
            -045import 
            org.apache.hadoop.conf.Configuration;
            -046import 
            org.apache.hadoop.conf.Configured;
            -047import org.apache.hadoop.fs.FileSystem;
            -048import org.apache.hadoop.fs.Path;
            -049import 
            org.apache.hadoop.hbase.HBaseConfiguration;
            -050import 
            org.apache.hadoop.hbase.TableName;
            -051import 
            org.apache.hadoop.hbase.backup.BackupAdmin;
            -052import 
            org.apache.hadoop.hbase.backup.BackupInfo;
            -053import 
            

            [26/51] [partial] hbase-site git commit: Published site at 7d3750bd9fc9747623549c242cc4171e224b3eaf.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html
            index 275ca84..b9a30c4 100644
            --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html
            +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.html
            @@ -1499,782 +1499,786 @@
             1491  final RpcController controller,
             1492  final IsInMaintenanceModeRequest 
            request) throws ServiceException {
             1493IsInMaintenanceModeResponse.Builder 
            response = IsInMaintenanceModeResponse.newBuilder();
            -1494
            response.setInMaintenanceMode(master.isInMaintenanceMode());
            -1495return response.build();
            -1496  }
            -1497
            -1498  @Override
            -1499  public UnassignRegionResponse 
            unassignRegion(RpcController controller,
            -1500  UnassignRegionRequest req) throws 
            ServiceException {
            -1501try {
            -1502  final byte [] regionName = 
            req.getRegion().getValue().toByteArray();
            -1503  RegionSpecifierType type = 
            req.getRegion().getType();
            -1504  final boolean force = 
            req.getForce();
            -1505  UnassignRegionResponse urr = 
            UnassignRegionResponse.newBuilder().build();
            -1506
            -1507  master.checkInitialized();
            -1508  if (type != 
            RegionSpecifierType.REGION_NAME) {
            -1509LOG.warn("unassignRegion 
            specifier type: expected: " + RegionSpecifierType.REGION_NAME
            -1510  + " actual: " + type);
            -1511  }
            -1512  PairRegionInfo, ServerName 
            pair =
            -1513
            MetaTableAccessor.getRegion(master.getConnection(), regionName);
            -1514  if 
            (Bytes.equals(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(),regionName))
             {
            -1515pair = new 
            Pair(RegionInfoBuilder.FIRST_META_REGIONINFO,
            -1516
            master.getMetaTableLocator().getMetaRegionLocation(master.getZooKeeper()));
            -1517  }
            -1518  if (pair == null) {
            -1519throw new 
            UnknownRegionException(Bytes.toString(regionName));
            -1520  }
            -1521
            -1522  RegionInfo hri = 
            pair.getFirst();
            -1523  if (master.cpHost != null) {
            -1524master.cpHost.preUnassign(hri, 
            force);
            -1525  }
            -1526  
            LOG.debug(master.getClientIdAuditPrefix() + " unassign " + 
            hri.getRegionNameAsString()
            -1527  + " in current location if it 
            is online and reassign.force=" + force);
            -1528  
            master.getAssignmentManager().unassign(hri);
            -1529  if (master.cpHost != null) {
            -1530master.cpHost.postUnassign(hri, 
            force);
            -1531  }
            -1532
            -1533  return urr;
            -1534} catch (IOException ioe) {
            -1535  throw new ServiceException(ioe);
            -1536}
            -1537  }
            -1538
            -1539  @Override
            -1540  public 
            ReportRegionStateTransitionResponse reportRegionStateTransition(RpcController 
            c,
            -1541  ReportRegionStateTransitionRequest 
            req) throws ServiceException {
            -1542try {
            -1543  master.checkServiceStarted();
            -1544  return 
            master.getAssignmentManager().reportRegionStateTransition(req);
            -1545} catch (IOException ioe) {
            -1546  throw new ServiceException(ioe);
            -1547}
            -1548  }
            -1549
            -1550  @Override
            -1551  public SetQuotaResponse 
            setQuota(RpcController c, SetQuotaRequest req)
            -1552  throws ServiceException {
            -1553try {
            -1554  master.checkInitialized();
            -1555  return 
            master.getMasterQuotaManager().setQuota(req);
            -1556} catch (Exception e) {
            -1557  throw new ServiceException(e);
            -1558}
            -1559  }
            -1560
            -1561  @Override
            -1562  public 
            MajorCompactionTimestampResponse getLastMajorCompactionTimestamp(RpcController 
            controller,
            -1563  MajorCompactionTimestampRequest 
            request) throws ServiceException {
            -1564
            MajorCompactionTimestampResponse.Builder response =
            -1565
            MajorCompactionTimestampResponse.newBuilder();
            -1566try {
            -1567  master.checkInitialized();
            -1568  
            response.setCompactionTimestamp(master.getLastMajorCompactionTimestamp(ProtobufUtil
            -1569  
            .toTableName(request.getTableName(;
            -1570} catch (IOException e) {
            -1571  throw new ServiceException(e);
            -1572}
            -1573return response.build();
            -1574  }
            -1575
            -1576  @Override
            -1577  public 
            MajorCompactionTimestampResponse getLastMajorCompactionTimestampForRegion(
            -1578  RpcController controller, 
            MajorCompactionTimestampForRegionRequest request)
            -1579  throws ServiceException {
            -1580
            MajorCompactionTimestampResponse.Builder response =
            -1581
            MajorCompactionTimestampResponse.newBuilder();
            -1582try {
            -1583  master.checkInitialized();
            -1584  
            response.setCompactionTimestamp(master.getLastMajorCompactionTimestampForRegion(request
            -1585  
            .getRegion().getValue().toByteArray()));
            -1586} catch (IOException e) {
            -1587  throw new ServiceException(e);
            -1588}
            -1589return response.build();
            -1590  }
            -1591
            -1592  /**
            -1593   * 

            [26/51] [partial] hbase-site git commit: Published site at 997747076d8ec0b4346d7cb99c4b0667a7c14905.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
            index 3da432b..d30fa8f 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
            @@ -928,7690 +928,7698 @@
             920  CollectionHStore stores = 
            this.stores.values();
             921  try {
             922// update the stores that we are 
            replaying
            -923
            stores.forEach(HStore::startReplayingFromWAL);
            -924// Recover any edits if 
            available.
            -925maxSeqId = Math.max(maxSeqId,
            -926  
            replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
            status));
            -927// Make sure mvcc is up to max.
            -928this.mvcc.advanceTo(maxSeqId);
            -929  } finally {
            -930// update the stores that we are 
            done replaying
            -931
            stores.forEach(HStore::stopReplayingFromWAL);
            -932  }
            -933}
            -934this.lastReplayedOpenRegionSeqId = 
            maxSeqId;
            -935
            -936
            this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
            -937this.writestate.flushRequested = 
            false;
            -938this.writestate.compacting.set(0);
            -939
            -940if (this.writestate.writesEnabled) 
            {
            -941  // Remove temporary data left over 
            from old regions
            -942  status.setStatus("Cleaning up 
            temporary data from old regions");
            -943  fs.cleanupTempDir();
            -944}
            -945
            -946if (this.writestate.writesEnabled) 
            {
            -947  status.setStatus("Cleaning up 
            detritus from prior splits");
            -948  // Get rid of any splits or merges 
            that were lost in-progress.  Clean out
            -949  // these directories here on open.  
            We may be opening a region that was
            -950  // being split but we crashed in 
            the middle of it all.
            -951  fs.cleanupAnySplitDetritus();
            -952  fs.cleanupMergesDir();
            -953}
            -954
            -955// Initialize split policy
            -956this.splitPolicy = 
            RegionSplitPolicy.create(this, conf);
            -957
            -958// Initialize flush policy
            -959this.flushPolicy = 
            FlushPolicyFactory.create(this, conf);
            -960
            -961long lastFlushTime = 
            EnvironmentEdgeManager.currentTime();
            -962for (HStore store: stores.values()) 
            {
            -963  
            this.lastStoreFlushTimeMap.put(store, lastFlushTime);
            -964}
            -965
            -966// Use maximum of log sequenceid or 
            that which was found in stores
            -967// (particularly if no recovered 
            edits, seqid will be -1).
            -968long maxSeqIdFromFile =
            -969  
            WALSplitter.getMaxRegionSequenceId(fs.getFileSystem(), fs.getRegionDir());
            -970long nextSeqId = Math.max(maxSeqId, 
            maxSeqIdFromFile) + 1;
            -971if (writestate.writesEnabled) {
            -972  
            WALSplitter.writeRegionSequenceIdFile(fs.getFileSystem(), fs.getRegionDir(), 
            nextSeqId - 1);
            -973}
            -974
            -975LOG.info("Opened {}; next 
            sequenceid={}", this.getRegionInfo().getShortNameToLog(), nextSeqId);
            -976
            -977// A region can be reopened if failed 
            a split; reset flags
            -978this.closing.set(false);
            -979this.closed.set(false);
            -980
            -981if (coprocessorHost != null) {
            -982  status.setStatus("Running 
            coprocessor post-open hooks");
            -983  coprocessorHost.postOpen();
            -984}
            +923LOG.debug("replaying wal for " + 
            this.getRegionInfo().getEncodedName());
            +924
            stores.forEach(HStore::startReplayingFromWAL);
            +925// Recover any edits if 
            available.
            +926maxSeqId = Math.max(maxSeqId,
            +927  
            replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
            status));
            +928// Make sure mvcc is up to max.
            +929this.mvcc.advanceTo(maxSeqId);
            +930  } finally {
            +931LOG.debug("stopping wal replay 
            for " + this.getRegionInfo().getEncodedName());
            +932// update the stores that we are 
            done replaying
            +933
            stores.forEach(HStore::stopReplayingFromWAL);
            +934  }
            +935}
            +936this.lastReplayedOpenRegionSeqId = 
            maxSeqId;
            +937
            +938
            this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
            +939this.writestate.flushRequested = 
            false;
            +940this.writestate.compacting.set(0);
            +941
            +942if (this.writestate.writesEnabled) 
            {
            +943  LOG.debug("Cleaning up temporary 
            data for " + this.getRegionInfo().getEncodedName());
            +944  // Remove temporary data left over 
            from old regions
            +945  status.setStatus("Cleaning up 
            temporary data from old regions");
            +946  fs.cleanupTempDir();
            +947}
            +948
            +949if (this.writestate.writesEnabled) 
            {
            +950  status.setStatus("Cleaning up 
            detritus from prior splits");
            +951  // Get rid of any splits or merges 
            that were lost in-progress.  Clean out
            +952  // these directories here on 

            [26/51] [partial] hbase-site git commit: Published site at f3d1c021de2264301f68eadb9ef126ff83d7ef53.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
            index 9644187..b979909 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
            @@ -1199,2602 +1199,2599 @@
             1191ClusterStatusProtos.ServerLoad sl = 
            buildServerLoad(reportStartTime, reportEndTime);
             1192try {
             1193  RegionServerReportRequest.Builder 
            request = RegionServerReportRequest.newBuilder();
            -1194  ServerName sn = 
            ServerName.parseVersionedServerName(this.serverName.getVersionedBytes());
            -1195  
            request.setServer(ProtobufUtil.toServerName(sn));
            -1196  request.setLoad(sl);
            -1197  rss.regionServerReport(null, 
            request.build());
            -1198} catch (ServiceException se) {
            -1199  IOException ioe = 
            ProtobufUtil.getRemoteException(se);
            -1200  if (ioe instanceof 
            YouAreDeadException) {
            -1201// This will be caught and 
            handled as a fatal error in run()
            -1202throw ioe;
            -1203  }
            -1204  if (rssStub == rss) {
            -1205rssStub = null;
            -1206  }
            -1207  // Couldn't connect to the master, 
            get location from zk and reconnect
            -1208  // Method blocks until new master 
            is found or we are stopped
            -1209  
            createRegionServerStatusStub(true);
            -1210}
            -1211  }
            -1212
            -1213  /**
            -1214   * Reports the given map of Regions 
            and their size on the filesystem to the active Master.
            -1215   *
            -1216   * @param regionSizeStore The store 
            containing region sizes
            -1217   * @return false if 
            FileSystemUtilizationChore should pause reporting to master. true otherwise
            -1218   */
            -1219  public boolean 
            reportRegionSizesForQuotas(RegionSizeStore regionSizeStore) {
            -1220
            RegionServerStatusService.BlockingInterface rss = rssStub;
            -1221if (rss == null) {
            -1222  // the current server could be 
            stopping.
            -1223  LOG.trace("Skipping Region size 
            report to HMaster as stub is null");
            -1224  return true;
            -1225}
            -1226try {
            -1227  buildReportAndSend(rss, 
            regionSizeStore);
            -1228} catch (ServiceException se) {
            -1229  IOException ioe = 
            ProtobufUtil.getRemoteException(se);
            -1230  if (ioe instanceof 
            PleaseHoldException) {
            -1231LOG.trace("Failed to report 
            region sizes to Master because it is initializing."
            -1232+ " This will be retried.", 
            ioe);
            -1233// The Master is coming up. Will 
            retry the report later. Avoid re-creating the stub.
            -1234return true;
            -1235  }
            -1236  if (rssStub == rss) {
            -1237rssStub = null;
            -1238  }
            -1239  
            createRegionServerStatusStub(true);
            -1240  if (ioe instanceof 
            DoNotRetryIOException) {
            -1241DoNotRetryIOException 
            doNotRetryEx = (DoNotRetryIOException) ioe;
            -1242if (doNotRetryEx.getCause() != 
            null) {
            -1243  Throwable t = 
            doNotRetryEx.getCause();
            -1244  if (t instanceof 
            UnsupportedOperationException) {
            -1245LOG.debug("master doesn't 
            support ReportRegionSpaceUse, pause before retrying");
            -1246return false;
            -1247  }
            -1248}
            -1249  }
            -1250  LOG.debug("Failed to report region 
            sizes to Master. This will be retried.", ioe);
            -1251}
            -1252return true;
            -1253  }
            -1254
            -1255  /**
            -1256   * Builds the region size report and 
            sends it to the master. Upon successful sending of the
            -1257   * report, the region sizes that were 
            sent are marked as sent.
            -1258   *
            -1259   * @param rss The stub to send to the 
            Master
            -1260   * @param regionSizeStore The store 
            containing region sizes
            -1261   */
            -1262  void 
            buildReportAndSend(RegionServerStatusService.BlockingInterface rss,
            -1263  RegionSizeStore regionSizeStore) 
            throws ServiceException {
            -1264RegionSpaceUseReportRequest request 
            =
            -1265
            buildRegionSpaceUseReportRequest(Objects.requireNonNull(regionSizeStore));
            -1266rss.reportRegionSpaceUse(null, 
            request);
            -1267// Record the number of size reports 
            sent
            -1268if (metricsRegionServer != null) {
            -1269  
            metricsRegionServer.incrementNumRegionSizeReportsSent(regionSizeStore.size());
            -1270}
            -1271  }
            -1272
            -1273  /**
            -1274   * Builds a {@link 
            RegionSpaceUseReportRequest} protobuf message from the region size map.
            -1275   *
            -1276   * @param regionSizeStore The size in 
            bytes of regions
            -1277   * @return The corresponding protocol 
            buffer message.
            -1278   */
            -1279  RegionSpaceUseReportRequest 
            buildRegionSpaceUseReportRequest(RegionSizeStore regionSizes) {
            -1280RegionSpaceUseReportRequest.Builder 
            request = RegionSpaceUseReportRequest.newBuilder();
            -1281for 

            [26/51] [partial] hbase-site git commit: Published site at cf529f18a9959589fa635f78df4840472526ea2c.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
            index 8b2674f..274eb54 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
            @@ -262,1413 +262,1417 @@
             254});
             255  }
             256
            -257  public void preModifyNamespace(final 
            NamespaceDescriptor ns) throws IOException {
            -258
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            -259  @Override
            -260  public void call(MasterObserver 
            observer) throws IOException {
            -261observer.preModifyNamespace(this, 
            ns);
            -262  }
            -263});
            -264  }
            -265
            -266  public void postModifyNamespace(final 
            NamespaceDescriptor ns) throws IOException {
            -267
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            -268  @Override
            -269  public void call(MasterObserver 
            observer) throws IOException {
            -270
            observer.postModifyNamespace(this, ns);
            -271  }
            -272});
            -273  }
            -274
            -275  public void 
            preGetNamespaceDescriptor(final String namespaceName)
            -276  throws IOException {
            -277
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            -278  @Override
            -279  public void call(MasterObserver 
            observer) throws IOException {
            -280
            observer.preGetNamespaceDescriptor(this, namespaceName);
            -281  }
            -282});
            -283  }
            -284
            -285  public void 
            postGetNamespaceDescriptor(final NamespaceDescriptor ns)
            -286  throws IOException {
            -287
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            -288  @Override
            -289  public void call(MasterObserver 
            observer) throws IOException {
            -290
            observer.postGetNamespaceDescriptor(this, ns);
            -291  }
            -292});
            -293  }
            -294
            -295  public void 
            preListNamespaceDescriptors(final ListNamespaceDescriptor 
            descriptors)
            -296  throws IOException {
            -297
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            -298  @Override
            -299  public void call(MasterObserver 
            observer) throws IOException {
            -300
            observer.preListNamespaceDescriptors(this, descriptors);
            -301  }
            -302});
            -303  }
            -304
            -305  public void 
            postListNamespaceDescriptors(final ListNamespaceDescriptor 
            descriptors)
            -306  throws IOException {
            -307
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            -308  @Override
            -309  public void call(MasterObserver 
            observer) throws IOException {
            -310
            observer.postListNamespaceDescriptors(this, descriptors);
            -311  }
            -312});
            -313  }
            -314
            -315  /* Implementation of hooks for invoking 
            MasterObservers */
            +257  public void preModifyNamespace(final 
            NamespaceDescriptor currentNsDescriptor,
            +258final NamespaceDescriptor 
            newNsDescriptor) throws IOException {
            +259
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            +260  @Override
            +261  public void call(MasterObserver 
            observer) throws IOException {
            +262observer.preModifyNamespace(this, 
            currentNsDescriptor, newNsDescriptor);
            +263  }
            +264});
            +265  }
            +266
            +267  public void postModifyNamespace(final 
            NamespaceDescriptor oldNsDescriptor,
            +268final NamespaceDescriptor 
            currentNsDescriptor) throws IOException {
            +269
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            +270  @Override
            +271  public void call(MasterObserver 
            observer) throws IOException {
            +272
            observer.postModifyNamespace(this, oldNsDescriptor, currentNsDescriptor);
            +273  }
            +274});
            +275  }
            +276
            +277  public void 
            preGetNamespaceDescriptor(final String namespaceName)
            +278  throws IOException {
            +279
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            +280  @Override
            +281  public void call(MasterObserver 
            observer) throws IOException {
            +282
            observer.preGetNamespaceDescriptor(this, namespaceName);
            +283  }
            +284});
            +285  }
            +286
            +287  public void 
            postGetNamespaceDescriptor(final NamespaceDescriptor ns)
            +288  throws IOException {
            +289
            execOperation(coprocEnvironments.isEmpty() ? null : new 
            MasterObserverOperation() {
            +290  @Override
            +291  public void call(MasterObserver 
            observer) throws IOException {
            +292
            observer.postGetNamespaceDescriptor(this, ns);
            +293  }
            +294});
            +295  }
            +296
            +297  public void 
            preListNamespaceDescriptors(final ListNamespaceDescriptor 
            descriptors)
            +298  throws 

            [26/51] [partial] hbase-site git commit: Published site at 021f66d11d2cbb7308308093e29e69d6e7661ee9.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
            index 4a879bb..7d27402 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
            @@ -300,7 +300,7 @@
             292  private MapString, 
            com.google.protobuf.Service coprocessorServiceHandlers = 
            Maps.newHashMap();
             293
             294  // Track data size in all memstores
            -295  private final MemStoreSizing 
            memStoreSize = new MemStoreSizing();
            +295  private final MemStoreSizing 
            memStoreSizing = new ThreadSafeMemStoreSizing();
             296  private final RegionServicesForStores 
            regionServicesForStores = new RegionServicesForStores(this);
             297
             298  // Debug possible data loss due to WAL 
            off
            @@ -1218,7389 +1218,7399 @@
             1210   * Increase the size of mem store in 
            this region and the size of global mem
             1211   * store
             1212   */
            -1213  public void 
            incMemStoreSize(MemStoreSize memStoreSize) {
            -1214if (this.rsAccounting != null) {
            -1215  
            rsAccounting.incGlobalMemStoreSize(memStoreSize);
            -1216}
            -1217long dataSize;
            -1218synchronized (this.memStoreSize) {
            -1219  
            this.memStoreSize.incMemStoreSize(memStoreSize);
            -1220  dataSize = 
            this.memStoreSize.getDataSize();
            -1221}
            -1222
            checkNegativeMemStoreDataSize(dataSize, memStoreSize.getDataSize());
            -1223  }
            -1224
            -1225  public void 
            decrMemStoreSize(MemStoreSize memStoreSize) {
            -1226if (this.rsAccounting != null) {
            -1227  
            rsAccounting.decGlobalMemStoreSize(memStoreSize);
            -1228}
            -1229long size;
            -1230synchronized (this.memStoreSize) {
            -1231  
            this.memStoreSize.decMemStoreSize(memStoreSize);
            -1232  size = 
            this.memStoreSize.getDataSize();
            +1213  void incMemStoreSize(MemStoreSize mss) 
            {
            +1214incMemStoreSize(mss.getDataSize(), 
            mss.getHeapSize(), mss.getOffHeapSize());
            +1215  }
            +1216
            +1217  void incMemStoreSize(long 
            dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) {
            +1218if (this.rsAccounting != null) {
            +1219  
            rsAccounting.incGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, 
            offHeapSizeDelta);
            +1220}
            +1221long dataSize =
            +1222
            this.memStoreSizing.incMemStoreSize(dataSizeDelta, heapSizeDelta, 
            offHeapSizeDelta);
            +1223
            checkNegativeMemStoreDataSize(dataSize, dataSizeDelta);
            +1224  }
            +1225
            +1226  void decrMemStoreSize(MemStoreSize 
            mss) {
            +1227decrMemStoreSize(mss.getDataSize(), 
            mss.getHeapSize(), mss.getOffHeapSize());
            +1228  }
            +1229
            +1230  void decrMemStoreSize(long 
            dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) {
            +1231if (this.rsAccounting != null) {
            +1232  
            rsAccounting.decGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, 
            offHeapSizeDelta);
             1233}
            -1234checkNegativeMemStoreDataSize(size, 
            -memStoreSize.getDataSize());
            -1235  }
            -1236
            -1237  private void 
            checkNegativeMemStoreDataSize(long memStoreDataSize, long delta) {
            -1238// This is extremely bad if we make 
            memStoreSize negative. Log as much info on the offending
            -1239// caller as possible. (memStoreSize 
            might be a negative value already -- freeing memory)
            -1240if (memStoreDataSize  0) {
            -1241  LOG.error("Asked to modify this 
            region's (" + this.toString()
            -1242  + ") memStoreSize to a 
            negative value which is incorrect. Current memStoreSize="
            -1243  + (memStoreDataSize - delta) + 
            ", delta=" + delta, new Exception());
            -1244}
            -1245  }
            -1246
            -1247  @Override
            -1248  public RegionInfo getRegionInfo() {
            -1249return this.fs.getRegionInfo();
            -1250  }
            -1251
            -1252  /**
            -1253   * @return Instance of {@link 
            RegionServerServices} used by this HRegion.
            -1254   * Can be null.
            -1255   */
            -1256  RegionServerServices 
            getRegionServerServices() {
            -1257return this.rsServices;
            -1258  }
            -1259
            -1260  @Override
            -1261  public long getReadRequestsCount() {
            -1262return readRequestsCount.sum();
            -1263  }
            -1264
            -1265  @Override
            -1266  public long 
            getFilteredReadRequestsCount() {
            -1267return 
            filteredReadRequestsCount.sum();
            -1268  }
            -1269
            -1270  @Override
            -1271  public long getWriteRequestsCount() 
            {
            -1272return writeRequestsCount.sum();
            -1273  }
            -1274
            -1275  @Override
            -1276  public long getMemStoreDataSize() {
            -1277return memStoreSize.getDataSize();
            -1278  }
            -1279
            -1280  @Override
            -1281  public long getMemStoreHeapSize() {
            -1282return memStoreSize.getHeapSize();
            -1283  }
            -1284
            -1285  @Override
            -1286  public long getMemStoreOffHeapSize() 
            {
            -1287return 
            memStoreSize.getOffHeapSize();
            -1288  }
            -1289
            -1290  /** @return store services for this 
            region, to access services required by store level needs */
            -1291  public 

            [26/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html
            --
            diff --git 
            a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html
             
            b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html
            index 2510283..418c60c 100644
            --- 
            a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html
            +++ 
            b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html
            @@ -77,77 +77,77 @@
             069import 
            org.apache.hadoop.hbase.client.RowMutations;
             070import 
            org.apache.hadoop.hbase.client.Scan;
             071import 
            org.apache.hadoop.hbase.client.Table;
            -072import 
            org.apache.hadoop.hbase.filter.BinaryComparator;
            -073import 
            org.apache.hadoop.hbase.filter.Filter;
            -074import 
            org.apache.hadoop.hbase.filter.FilterAllFilter;
            -075import 
            org.apache.hadoop.hbase.filter.FilterList;
            -076import 
            org.apache.hadoop.hbase.filter.PageFilter;
            -077import 
            org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
            -078import 
            org.apache.hadoop.hbase.filter.WhileMatchFilter;
            -079import 
            org.apache.hadoop.hbase.io.compress.Compression;
            -080import 
            org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
            -081import 
            org.apache.hadoop.hbase.io.hfile.RandomDistribution;
            -082import 
            org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
            -083import 
            org.apache.hadoop.hbase.regionserver.BloomType;
            -084import 
            org.apache.hadoop.hbase.regionserver.CompactingMemStore;
            -085import 
            org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration;
            -086import 
            org.apache.hadoop.hbase.trace.SpanReceiverHost;
            -087import 
            org.apache.hadoop.hbase.trace.TraceUtil;
            -088import 
            org.apache.hadoop.hbase.util.ByteArrayHashKey;
            -089import 
            org.apache.hadoop.hbase.util.Bytes;
            -090import 
            org.apache.hadoop.hbase.util.Hash;
            -091import 
            org.apache.hadoop.hbase.util.MurmurHash;
            -092import 
            org.apache.hadoop.hbase.util.Pair;
            -093import 
            org.apache.hadoop.hbase.util.YammerHistogramUtils;
            -094import 
            org.apache.hadoop.io.LongWritable;
            -095import org.apache.hadoop.io.Text;
            -096import org.apache.hadoop.mapreduce.Job;
            -097import 
            org.apache.hadoop.mapreduce.Mapper;
            -098import 
            org.apache.hadoop.mapreduce.lib.input.NLineInputFormat;
            -099import 
            org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
            -100import 
            org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer;
            -101import org.apache.hadoop.util.Tool;
            -102import 
            org.apache.hadoop.util.ToolRunner;
            -103import 
            org.apache.htrace.core.ProbabilitySampler;
            -104import org.apache.htrace.core.Sampler;
            -105import 
            org.apache.htrace.core.TraceScope;
            -106import 
            org.apache.yetus.audience.InterfaceAudience;
            -107import org.slf4j.Logger;
            -108import org.slf4j.LoggerFactory;
            -109import 
            org.apache.hbase.thirdparty.com.google.common.base.MoreObjects;
            -110import 
            org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
            -111
            -112/**
            -113 * Script used evaluating HBase 
            performance and scalability.  Runs a HBase
            -114 * client that steps through one of a set 
            of hardcoded tests or 'experiments'
            -115 * (e.g. a random reads test, a random 
            writes test, etc.). Pass on the
            -116 * command-line which test to run and how 
            many clients are participating in
            -117 * this experiment. Run {@code 
            PerformanceEvaluation --help} to obtain usage.
            -118 *
            -119 * pThis class sets up and runs 
            the evaluation programs described in
            -120 * Section 7, iPerformance 
            Evaluation/i, of the a
            -121 * 
            href="http://labs.google.com/papers/bigtable.html"Bigtable/a;
            -122 * paper, pages 8-10.
            -123 *
            -124 * pBy default, runs as a 
            mapreduce job where each mapper runs a single test
            -125 * client. Can also run as a 
            non-mapreduce, multithreaded application by
            -126 * specifying {@code --nomapred}. Each 
            client does about 1GB of data, unless
            -127 * specified otherwise.
            -128 */
            -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
            -130public class PerformanceEvaluation 
            extends Configured implements Tool {
            -131  static final String RANDOM_SEEK_SCAN = 
            "randomSeekScan";
            -132  static final String RANDOM_READ = 
            "randomRead";
            -133  private static final Logger LOG = 
            LoggerFactory.getLogger(PerformanceEvaluation.class.getName());
            -134  private static final ObjectMapper 
            MAPPER = new ObjectMapper();
            -135  static {
            -136
            MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true);
            -137  }
            -138
            -139  public static final String TABLE_NAME = 
            "TestTable";
            -140  public static final byte[] FAMILY_NAME 
            = Bytes.toBytes("info");
            -141  public static final byte [] COLUMN_ZERO 
            = Bytes.toBytes("" + 0);
            -142  public static final byte [] 
            QUALIFIER_NAME = COLUMN_ZERO;
            +072import 
            org.apache.hadoop.hbase.client.metrics.ScanMetrics;
            +073import 
            org.apache.hadoop.hbase.filter.BinaryComparator;
            +074import 
            org.apache.hadoop.hbase.filter.Filter;
            +075import 
            

            [26/51] [partial] hbase-site git commit: Published site at 87f5b5f3411d96c31b4cb61b9a57ced22be91d1f.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html
            index e1bc325..63e7421 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.OnlineEntry.html
            @@ -66,5125 +66,5224 @@
             058import 
            java.util.concurrent.TimeoutException;
             059import 
            java.util.concurrent.atomic.AtomicBoolean;
             060import 
            java.util.concurrent.atomic.AtomicInteger;
            -061import org.apache.commons.io.IOUtils;
            -062import 
            org.apache.commons.lang3.RandomStringUtils;
            -063import 
            org.apache.commons.lang3.StringUtils;
            -064import 
            org.apache.hadoop.conf.Configuration;
            -065import 
            org.apache.hadoop.conf.Configured;
            -066import 
            org.apache.hadoop.fs.FSDataOutputStream;
            -067import org.apache.hadoop.fs.FileStatus;
            -068import org.apache.hadoop.fs.FileSystem;
            -069import org.apache.hadoop.fs.Path;
            -070import 
            org.apache.hadoop.fs.permission.FsAction;
            -071import 
            org.apache.hadoop.fs.permission.FsPermission;
            -072import 
            org.apache.hadoop.hbase.Abortable;
            -073import org.apache.hadoop.hbase.Cell;
            -074import 
            org.apache.hadoop.hbase.CellUtil;
            -075import 
            org.apache.hadoop.hbase.ClusterMetrics;
            -076import 
            org.apache.hadoop.hbase.ClusterMetrics.Option;
            -077import 
            org.apache.hadoop.hbase.HBaseConfiguration;
            -078import 
            org.apache.hadoop.hbase.HBaseInterfaceAudience;
            -079import 
            org.apache.hadoop.hbase.HConstants;
            -080import 
            org.apache.hadoop.hbase.HRegionInfo;
            -081import 
            org.apache.hadoop.hbase.HRegionLocation;
            -082import 
            org.apache.hadoop.hbase.KeyValue;
            -083import 
            org.apache.hadoop.hbase.MasterNotRunningException;
            -084import 
            org.apache.hadoop.hbase.MetaTableAccessor;
            -085import 
            org.apache.hadoop.hbase.RegionLocations;
            -086import 
            org.apache.hadoop.hbase.ServerName;
            -087import 
            org.apache.hadoop.hbase.TableName;
            -088import 
            org.apache.hadoop.hbase.ZooKeeperConnectionException;
            -089import 
            org.apache.hadoop.hbase.client.Admin;
            -090import 
            org.apache.hadoop.hbase.client.ClusterConnection;
            -091import 
            org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
            -092import 
            org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
            -093import 
            org.apache.hadoop.hbase.client.Connection;
            -094import 
            org.apache.hadoop.hbase.client.ConnectionFactory;
            -095import 
            org.apache.hadoop.hbase.client.Delete;
            -096import 
            org.apache.hadoop.hbase.client.Get;
            -097import 
            org.apache.hadoop.hbase.client.Put;
            -098import 
            org.apache.hadoop.hbase.client.RegionInfo;
            -099import 
            org.apache.hadoop.hbase.client.RegionInfoBuilder;
            -100import 
            org.apache.hadoop.hbase.client.RegionReplicaUtil;
            -101import 
            org.apache.hadoop.hbase.client.Result;
            -102import 
            org.apache.hadoop.hbase.client.RowMutations;
            -103import 
            org.apache.hadoop.hbase.client.Table;
            -104import 
            org.apache.hadoop.hbase.client.TableDescriptor;
            -105import 
            org.apache.hadoop.hbase.client.TableDescriptorBuilder;
            -106import 
            org.apache.hadoop.hbase.client.TableState;
            -107import 
            org.apache.hadoop.hbase.io.FileLink;
            -108import 
            org.apache.hadoop.hbase.io.HFileLink;
            -109import 
            org.apache.hadoop.hbase.io.hfile.CacheConfig;
            -110import 
            org.apache.hadoop.hbase.io.hfile.HFile;
            -111import 
            org.apache.hadoop.hbase.log.HBaseMarkers;
            -112import 
            org.apache.hadoop.hbase.master.MasterFileSystem;
            -113import 
            org.apache.hadoop.hbase.master.RegionState;
            -114import 
            org.apache.hadoop.hbase.regionserver.HRegion;
            -115import 
            org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
            -116import 
            org.apache.hadoop.hbase.regionserver.StoreFileInfo;
            -117import 
            org.apache.hadoop.hbase.replication.ReplicationException;
            -118import 
            org.apache.hadoop.hbase.security.AccessDeniedException;
            -119import 
            org.apache.hadoop.hbase.security.UserProvider;
            -120import 
            org.apache.hadoop.hbase.util.Bytes.ByteArrayComparator;
            -121import 
            org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE;
            -122import 
            org.apache.hadoop.hbase.util.hbck.HFileCorruptionChecker;
            -123import 
            org.apache.hadoop.hbase.util.hbck.ReplicationChecker;
            -124import 
            org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandler;
            -125import 
            org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandlerImpl;
            -126import org.apache.hadoop.hbase.wal.WAL;
            -127import 
            org.apache.hadoop.hbase.wal.WALFactory;
            -128import 
            org.apache.hadoop.hbase.wal.WALSplitter;
            -129import 
            org.apache.hadoop.hbase.zookeeper.MetaTableLocator;
            -130import 
            org.apache.hadoop.hbase.zookeeper.ZKUtil;
            -131import 
            org.apache.hadoop.hbase.zookeeper.ZKWatcher;
            -132import 
            org.apache.hadoop.hbase.zookeeper.ZNodePaths;
            -133import 
            org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException;
            -134import 
            org.apache.hadoop.ipc.RemoteException;
            -135import 
            org.apache.hadoop.security.UserGroupInformation;
            

            [26/51] [partial] hbase-site git commit: Published site at 2912c953551bedbfbf30c32c156ed7bb187d54c3.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
            index cdb80a0..b442170 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
            @@ -126,227 +126,226 @@
             118  final long openSeqNum = -1;
             119
             120  // TODO: move under trace, now is 
            visible for debugging
            -121  LOG.info(String.format("Load 
            hbase:meta entry region=%s regionState=%s lastHost=%s regionLocation=%s",
            -122regionInfo, state, lastHost, 
            regionLocation));
            -123
            -124  
            visitor.visitRegionState(regionInfo, state, regionLocation, lastHost, 
            openSeqNum);
            -125}
            -126  }
            -127
            -128  public void 
            updateRegionLocation(RegionStates.RegionStateNode regionStateNode)
            -129  throws IOException {
            -130if 
            (regionStateNode.getRegionInfo().isMetaRegion()) {
            -131  
            updateMetaLocation(regionStateNode.getRegionInfo(), 
            regionStateNode.getRegionLocation());
            -132} else {
            -133  long openSeqNum = 
            regionStateNode.getState() == State.OPEN ?
            -134  regionStateNode.getOpenSeqNum() 
            : HConstants.NO_SEQNUM;
            -135  
            updateUserRegionLocation(regionStateNode.getRegionInfo(), 
            regionStateNode.getState(),
            -136  
            regionStateNode.getRegionLocation(), regionStateNode.getLastHost(), 
            openSeqNum,
            -137  
            regionStateNode.getProcedure().getProcId());
            -138}
            -139  }
            -140
            -141  private void updateMetaLocation(final 
            RegionInfo regionInfo, final ServerName serverName)
            -142  throws IOException {
            -143try {
            -144  
            MetaTableLocator.setMetaLocation(master.getZooKeeper(), serverName,
            -145regionInfo.getReplicaId(), 
            State.OPEN);
            -146} catch (KeeperException e) {
            -147  throw new IOException(e);
            -148}
            -149  }
            -150
            -151  private void 
            updateUserRegionLocation(final RegionInfo regionInfo, final State state,
            -152  final ServerName regionLocation, 
            final ServerName lastHost, final long openSeqNum,
            -153  final long pid)
            -154  throws IOException {
            -155long time = 
            EnvironmentEdgeManager.currentTime();
            -156final int replicaId = 
            regionInfo.getReplicaId();
            -157final Put put = new 
            Put(MetaTableAccessor.getMetaKeyForRegion(regionInfo), time);
            -158MetaTableAccessor.addRegionInfo(put, 
            regionInfo);
            -159final StringBuilder info =
            -160  new 
            StringBuilder("pid=").append(pid).append(" updating hbase:meta row=")
            -161
            .append(regionInfo.getRegionNameAsString()).append(", 
            regionState=").append(state);
            -162if (openSeqNum = 0) {
            -163  Preconditions.checkArgument(state 
            == State.OPEN  regionLocation != null,
            -164  "Open region should be on a 
            server");
            -165  MetaTableAccessor.addLocation(put, 
            regionLocation, openSeqNum, replicaId);
            -166  // only update replication barrier 
            for default replica
            -167  if (regionInfo.getReplicaId() == 
            RegionInfo.DEFAULT_REPLICA_ID 
            -168
            hasGlobalReplicationScope(regionInfo.getTable())) {
            -169
            MetaTableAccessor.addReplicationBarrier(put, openSeqNum);
            -170  }
            -171  info.append(", 
            openSeqNum=").append(openSeqNum);
            -172  info.append(", 
            regionLocation=").append(regionLocation);
            -173} else if (regionLocation != null 
             !regionLocation.equals(lastHost)) {
            -174  // Ideally, if no regionLocation, 
            write null to the hbase:meta but this will confuse clients
            -175  // currently; they want a server to 
            hit. TODO: Make clients wait if no location.
            -176  
            put.add(CellBuilderFactory.create(CellBuilderType.SHALLOW_COPY)
            -177  .setRow(put.getRow())
            -178  
            .setFamily(HConstants.CATALOG_FAMILY)
            -179  
            .setQualifier(getServerNameColumn(replicaId))
            -180  
            .setTimestamp(put.getTimestamp())
            -181  .setType(Cell.Type.Put)
            -182  
            .setValue(Bytes.toBytes(regionLocation.getServerName()))
            -183  .build());
            -184  info.append(", 
            regionLocation=").append(regionLocation);
            -185}
            -186
            put.add(CellBuilderFactory.create(CellBuilderType.SHALLOW_COPY)
            -187.setRow(put.getRow())
            -188
            .setFamily(HConstants.CATALOG_FAMILY)
            -189
            .setQualifier(getStateColumn(replicaId))
            -190
            .setTimestamp(put.getTimestamp())
            -191.setType(Cell.Type.Put)
            -192
            .setValue(Bytes.toBytes(state.name()))
            -193.build());
            -194LOG.info(info.toString());
            -195updateRegionLocation(regionInfo, 
            state, put);
            -196  }
            -197
            -198  private void 
            updateRegionLocation(RegionInfo regionInfo, State state, Put put)
            -199  throws IOException {
            -200try (Table table = 
            master.getConnection().getTable(TableName.META_TABLE_NAME)) {
            -201  

            [26/51] [partial] hbase-site git commit: Published site at 2a2258656b2fcd92b967131b6c1f037363553bc4.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/wal/WALPrettyPrinter.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/wal/WALPrettyPrinter.html 
            b/devapidocs/org/apache/hadoop/hbase/wal/WALPrettyPrinter.html
            index cc33120..3cc25e3 100644
            --- a/devapidocs/org/apache/hadoop/hbase/wal/WALPrettyPrinter.html
            +++ b/devapidocs/org/apache/hadoop/hbase/wal/WALPrettyPrinter.html
            @@ -111,7 +111,7 @@ var activeTableTab = "activeTableTab";
             
             @InterfaceAudience.LimitedPrivate(value="Tools")
              @InterfaceStability.Evolving
            -public class WALPrettyPrinter
            +public class WALPrettyPrinter
             extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
             WALPrettyPrinter prints the contents of a given WAL with a 
            variety of
              options affecting formatting and extent of content.
            @@ -327,7 +327,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             outputValues
            -privateboolean outputValues
            +privateboolean outputValues
             
             
             
            @@ -336,7 +336,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             outputJSON
            -privateboolean outputJSON
            +privateboolean outputJSON
             
             
             
            @@ -345,7 +345,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             sequence
            -privatelong sequence
            +privatelong sequence
             
             
             
            @@ -354,7 +354,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             region
            -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String region
            +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String region
             
             
             
            @@ -363,7 +363,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             row
            -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String row
            +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String row
             
             
             
            @@ -372,7 +372,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             persistentOutput
            -privateboolean persistentOutput
            +privateboolean persistentOutput
             
             
             
            @@ -381,7 +381,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             firstTxn
            -privateboolean firstTxn
            +privateboolean firstTxn
             
             
             
            @@ -390,7 +390,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             out
            -privatehttps://docs.oracle.com/javase/8/docs/api/java/io/PrintStream.html?is-external=true;
             title="class or interface in java.io">PrintStream out
            +privatehttps://docs.oracle.com/javase/8/docs/api/java/io/PrintStream.html?is-external=true;
             title="class or interface in java.io">PrintStream out
             
             
             
            @@ -399,7 +399,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             MAPPER
            -private static finalcom.fasterxml.jackson.databind.ObjectMapper MAPPER
            +private static finalcom.fasterxml.jackson.databind.ObjectMapper MAPPER
             
             
             
            @@ -416,7 +416,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             WALPrettyPrinter
            -publicWALPrettyPrinter()
            +publicWALPrettyPrinter()
             Basic constructor that simply initializes values to 
            reasonable defaults.
             
             
            @@ -426,7 +426,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             WALPrettyPrinter
            -publicWALPrettyPrinter(booleanoutputValues,
            +publicWALPrettyPrinter(booleanoutputValues,
             booleanoutputJSON,
             longsequence,
             https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringregion,
            @@ -467,7 +467,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             enableValues
            -publicvoidenableValues()
            +publicvoidenableValues()
             turns value output on
             
             
            @@ -477,7 +477,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             disableValues
            -publicvoiddisableValues()
            +publicvoiddisableValues()
             turns value output off
             
             
            @@ -487,7 +487,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             enableJSON
            -publicvoidenableJSON()
            +publicvoidenableJSON()
             turns JSON output on
             
             
            @@ -497,7 +497,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             disableJSON
            -publicvoiddisableJSON()
            +publicvoiddisableJSON()
             turns JSON output off, and turns on "pretty strings" for 
            human consumption
             
             
            @@ -507,7 +507,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
             
             
             setSequenceFilter
            -publicvoidsetSequenceFilter(longsequence)
            +publicvoidsetSequenceFilter(longsequence)
             sets the region by which output will 

            [26/51] [partial] hbase-site git commit: Published site at e468b4022f76688851b3e0c34722f01a56bd624f.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html
            --
            diff --git 
            a/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html 
            b/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html
            index ecc21f4..54d784d 100644
            --- a/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html
            +++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html
            @@ -45,207 +45,208 @@
             037 * p
             038 * Note: Use of this filter overrides any 
            time range/time stamp
             039 * options specified using {@link 
            org.apache.hadoop.hbase.client.Get#setTimeRange(long, long)},
            -040 * {@link 
            org.apache.hadoop.hbase.client.Scan#setTimeRange(long, long)}, {@link 
            org.apache.hadoop.hbase.client.Get#setTimeStamp(long)},
            -041 * or {@link 
            org.apache.hadoop.hbase.client.Scan#setTimeStamp(long)}.
            -042 */
            -043@InterfaceAudience.Public
            -044public class TimestampsFilter extends 
            FilterBase {
            -045
            -046  private final boolean canHint;
            -047  TreeSetLong timestamps;
            -048  private static final int 
            MAX_LOG_TIMESTAMPS = 5;
            -049
            -050  // Used during scans to hint the scan 
            to stop early
            -051  // once the timestamps fall below the 
            minTimeStamp.
            -052  long minTimeStamp = Long.MAX_VALUE;
            -053
            -054  /**
            -055   * Constructor for filter that retains 
            only the specified timestamps in the list.
            -056   * @param timestamps
            -057   */
            -058  public 
            TimestampsFilter(ListLong timestamps) {
            -059this(timestamps, false);
            -060  }
            -061
            -062  /**
            -063   * Constructor for filter that retains 
            only those
            -064   * cells whose timestamp (version) is 
            in the specified
            -065   * list of timestamps.
            -066   *
            -067   * @param timestamps list of timestamps 
            that are wanted.
            -068   * @param canHint should the filter 
            provide a seek hint? This can skip
            -069   *past delete 
            tombstones, so it should only be used when that
            -070   *is not an issue ( no 
            deletes, or don't care if data
            -071   *becomes visible)
            -072   */
            -073  public 
            TimestampsFilter(ListLong timestamps, boolean canHint) {
            -074for (Long timestamp : timestamps) {
            -075  
            Preconditions.checkArgument(timestamp = 0, "must be positive %s", 
            timestamp);
            -076}
            -077this.canHint = canHint;
            -078this.timestamps = new 
            TreeSet(timestamps);
            -079init();
            -080  }
            -081
            -082  /**
            -083   * @return the list of timestamps
            -084   */
            -085  public ListLong getTimestamps() 
            {
            -086ListLong list = new 
            ArrayList(timestamps.size());
            -087list.addAll(timestamps);
            -088return list;
            -089  }
            -090
            -091  private void init() {
            -092if (this.timestamps.size()  0) 
            {
            -093  minTimeStamp = 
            this.timestamps.first();
            -094}
            -095  }
            -096
            -097  /**
            -098   * Gets the minimum timestamp requested 
            by filter.
            -099   * @return  minimum timestamp requested 
            by filter.
            -100   */
            -101  public long getMin() {
            -102return minTimeStamp;
            -103  }
            -104
            -105  @Override
            -106  public boolean filterRowKey(Cell cell) 
            throws IOException {
            -107// Impl in FilterBase might do 
            unnecessary copy for Off heap backed Cells.
            -108return false;
            -109  }
            -110
            -111  @Deprecated
            -112  @Override
            -113  public ReturnCode filterKeyValue(final 
            Cell c) {
            -114return filterCell(c);
            -115  }
            -116
            -117  @Override
            -118  public ReturnCode filterCell(final Cell 
            c) {
            -119if 
            (this.timestamps.contains(c.getTimestamp())) {
            -120  return ReturnCode.INCLUDE;
            -121} else if (c.getTimestamp()  
            minTimeStamp) {
            -122  // The remaining versions of this 
            column are guaranteed
            -123  // to be lesser than all of the 
            other values.
            -124  return ReturnCode.NEXT_COL;
            -125}
            -126return canHint ? 
            ReturnCode.SEEK_NEXT_USING_HINT : ReturnCode.SKIP;
            -127  }
            -128
            +040 * {@link 
            org.apache.hadoop.hbase.client.Scan#setTimeRange(long, long)},
            +041 * {@link 
            org.apache.hadoop.hbase.client.Get#setTimestamp(long)},
            +042 * or {@link 
            org.apache.hadoop.hbase.client.Scan#setTimestamp(long)}.
            +043 */
            +044@InterfaceAudience.Public
            +045public class TimestampsFilter extends 
            FilterBase {
            +046
            +047  private final boolean canHint;
            +048  TreeSetLong timestamps;
            +049  private static final int 
            MAX_LOG_TIMESTAMPS = 5;
            +050
            +051  // Used during scans to hint the scan 
            to stop early
            +052  // once the timestamps fall below the 
            minTimestamp.
            +053  long minTimestamp = Long.MAX_VALUE;
            +054
            +055  /**
            +056   * Constructor for filter that retains 
            only the specified timestamps in the list.
            +057   * @param timestamps
            +058   */
            +059  public 
            TimestampsFilter(ListLong timestamps) {
            +060this(timestamps, false);
            +061  }
            +062
            +063  /**
            +064   * Constructor for filter that retains 
            only those
            +065   * cells whose timestamp (version) is 
            in the specified
            +066   * list of timestamps.
            +067   *
            +068   * @param timestamps list of timestamps 
            that are wanted.
            +069   * @param canHint should the filter 
            provide a seek hint? This can skip
            +070 

            [26/51] [partial] hbase-site git commit: Published site at 64061f896fe21512504e3886a400759e88b519da.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            index 7eb7661..1d50582 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            @@ -104,13 +104,13 @@
             
             
             void
            -MasterServices.checkTableModifiable(TableNametableName)
            -Check table is modifiable; i.e.
            -
            +HMaster.checkTableModifiable(TableNametableName)
             
             
             void
            -HMaster.checkTableModifiable(TableNametableName)
            +MasterServices.checkTableModifiable(TableNametableName)
            +Check table is modifiable; i.e.
            +
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            index 25d874c..6194cbc 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            @@ -170,13 +170,13 @@
             
             
             void
            -MasterServices.checkTableModifiable(TableNametableName)
            -Check table is modifiable; i.e.
            -
            +HMaster.checkTableModifiable(TableNametableName)
             
             
             void
            -HMaster.checkTableModifiable(TableNametableName)
            +MasterServices.checkTableModifiable(TableNametableName)
            +Check table is modifiable; i.e.
            +
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            index 22b3482..067aa54 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            @@ -243,10 +243,6 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTag
            -PrivateCellUtil.getTags(Cellcell)
            -
            -
            -static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTag
             CellUtil.getTags(Cellcell)
             Deprecated.
             As of release 2.0.0, this 
            will be removed in HBase 3.0.0.
            @@ -254,6 +250,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            +
            +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTag
            +PrivateCellUtil.getTags(Cellcell)
            +
             
             static https://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true;
             title="class or interface in java.util">IteratorTag
             CellUtil.tagsIterator(byte[]tags,
            @@ -395,11 +395,6 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             static Cell
            -PrivateCellUtil.createCell(Cellcell,
            -  https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
            -
            -
            -static Cell
             CellUtil.createCell(Cellcell,
               https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
             Deprecated.
            @@ -407,6 +402,11 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            +
            +static Cell
            +PrivateCellUtil.createCell(Cellcell,
            +  https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
            +
             
             static byte[]
             TagUtil.fromList(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
            @@ -415,17 +415,17 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -ExtendedCellBuilder
            -ExtendedCellBuilderImpl.setTags(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
            -
            -
             RawCellBuilder
             RawCellBuilder.setTags(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
             
            -
            +
             ExtendedCellBuilder
             ExtendedCellBuilder.setTags(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
             
            +
            +ExtendedCellBuilder
            

            [26/51] [partial] hbase-site git commit: Published site at 4cb40e6d846ce1f28ffb40d388c9efb753197813.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            index f0cb8c2..5360e0a 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
            @@ -2052,119 +2052,119 @@ service.
             
             
             private TableName
            -RegionCoprocessorRpcChannel.table
            +SnapshotDescription.table
             
             
             private TableName
            -SnapshotDescription.table
            +RegionCoprocessorRpcChannel.table
             
             
             private TableName
            -HRegionLocator.tableName
            +RawAsyncTableImpl.tableName
             
             
             private TableName
            -ScannerCallableWithReplicas.tableName
            +RegionServerCallable.tableName
             
             
             protected TableName
            -ClientScanner.tableName
            +RegionAdminServiceCallable.tableName
             
             
             private TableName
            -AsyncClientScanner.tableName
            +BufferedMutatorImpl.tableName
             
             
             private TableName
            -AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.tableName
            +AsyncProcessTask.tableName
             
             
             private TableName
            -AsyncRpcRetryingCallerFactory.BatchCallerBuilder.tableName
            +AsyncProcessTask.Builder.tableName
             
             
             private TableName
            -RegionInfoBuilder.tableName
            +AsyncRequestFutureImpl.tableName
             
             
            -private TableName
            -RegionInfoBuilder.MutableRegionInfo.tableName
            +protected TableName
            +TableBuilderBase.tableName
             
             
             private TableName
            -RawAsyncTableImpl.tableName
            +AsyncBatchRpcRetryingCaller.tableName
             
             
             private TableName
            -RegionCoprocessorRpcChannelImpl.tableName
            +RegionInfoBuilder.tableName
             
             
             private TableName
            -AsyncTableRegionLocatorImpl.tableName
            +RegionInfoBuilder.MutableRegionInfo.tableName
             
             
            -protected TableName
            -RegionAdminServiceCallable.tableName
            +private TableName
            +HTable.tableName
             
             
             private TableName
            -HTable.tableName
            +TableState.tableName
             
             
            -private TableName
            -BufferedMutatorImpl.tableName
            +protected TableName
            +RpcRetryingCallerWithReadReplicas.tableName
             
             
            -private TableName
            -AsyncBatchRpcRetryingCaller.tableName
            +protected TableName
            +AsyncTableBuilderBase.tableName
             
             
             private TableName
            -BufferedMutatorParams.tableName
            +AsyncSingleRequestRpcRetryingCaller.tableName
             
             
             private TableName
            -HBaseAdmin.TableFuture.tableName
            +ScannerCallableWithReplicas.tableName
             
             
            -private TableName
            -AsyncRequestFutureImpl.tableName
            +protected TableName
            +RawAsyncHBaseAdmin.TableProcedureBiConsumer.tableName
             
             
             private TableName
            -AsyncProcessTask.tableName
            +AsyncTableRegionLocatorImpl.tableName
             
             
             private TableName
            -AsyncProcessTask.Builder.tableName
            +HBaseAdmin.TableFuture.tableName
             
             
            -protected TableName
            -RawAsyncHBaseAdmin.TableProcedureBiConsumer.tableName
            +private TableName
            +RegionCoprocessorRpcChannelImpl.tableName
             
             
            -private TableName
            -RegionServerCallable.tableName
            +protected TableName
            +ClientScanner.tableName
             
             
             private TableName
            -AsyncSingleRequestRpcRetryingCaller.tableName
            +BufferedMutatorParams.tableName
             
             
            -protected TableName
            -TableBuilderBase.tableName
            +private TableName
            +AsyncClientScanner.tableName
             
             
            -protected TableName
            -RpcRetryingCallerWithReadReplicas.tableName
            +private TableName
            +AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.tableName
             
             
            -protected TableName
            -AsyncTableBuilderBase.tableName
            +private TableName
            +AsyncRpcRetryingCallerFactory.BatchCallerBuilder.tableName
             
             
             private TableName
            -TableState.tableName
            +HRegionLocator.tableName
             
             
             
            @@ -2206,83 +2206,83 @@ service.
             
             
             TableName
            -AsyncTable.getName()
            -Gets the fully qualified table name instance of this 
            table.
            -
            +RawAsyncTableImpl.getName()
             
             
             TableName
            -Table.getName()
            +RegionLocator.getName()
             Gets the fully qualified table name instance of this 
            table.
             
             
             
             TableName
            -HRegionLocator.getName()
            +BufferedMutatorImpl.getName()
             
             
             TableName
            -AsyncTableRegionLocator.getName()
            -Gets the fully qualified table name instance of the table 
            whose region we want to locate.
            +BufferedMutator.getName()
            +Gets the fully qualified table name instance of the table 
            that this BufferedMutator writes to.
             
             
             
             TableName
            -AsyncTableImpl.getName()
            +HTable.getName()
             
             
             TableName
            -RawAsyncTableImpl.getName()
            +AsyncBufferedMutator.getName()
            +Gets the fully qualified table name instance of the table 
            that this
            + AsyncBufferedMutator writes to.
            +
             
             
             TableName
            -AsyncTableRegionLocatorImpl.getName()
            +Table.getName()
            +Gets the fully qualified table name instance of this 
            table.
            +
             
             
             TableName
            -BufferedMutator.getName()
            -Gets the fully qualified table name instance of the table 
            that this BufferedMutator writes to.
            -
            +AsyncTableImpl.getName()
             
             
             TableName
            -RegionLocator.getName()
            +AsyncTable.getName()
             Gets the fully qualified table name instance of this 
            table.
             
             
             
             TableName
            -AsyncBufferedMutatorImpl.getName()
            +AsyncTableRegionLocatorImpl.getName()
             
             
             TableName
            -HTable.getName()
            +AsyncTableRegionLocator.getName()
            +Gets the fully qualified table 

            [26/51] [partial] hbase-site git commit: Published site at 8ab7b20f48951d77945181024f5e15842bc253c4.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
            index ecf500c..0cd5a4e 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
            @@ -238,8355 +238,8368 @@
             230  public static final String 
            HBASE_MAX_CELL_SIZE_KEY = "hbase.server.keyvalue.maxsize";
             231  public static final int 
            DEFAULT_MAX_CELL_SIZE = 10485760;
             232
            -233  public static final String 
            HBASE_REGIONSERVER_MINIBATCH_SIZE =
            -234  
            "hbase.regionserver.minibatch.size";
            -235  public static final int 
            DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2;
            -236
            -237  /**
            -238   * This is the global default value for 
            durability. All tables/mutations not
            -239   * defining a durability or using 
            USE_DEFAULT will default to this value.
            -240   */
            -241  private static final Durability 
            DEFAULT_DURABILITY = Durability.SYNC_WAL;
            +233  /**
            +234   * This is the global default value for 
            durability. All tables/mutations not
            +235   * defining a durability or using 
            USE_DEFAULT will default to this value.
            +236   */
            +237  private static final Durability 
            DEFAULT_DURABILITY = Durability.SYNC_WAL;
            +238
            +239  public static final String 
            HBASE_REGIONSERVER_MINIBATCH_SIZE =
            +240  
            "hbase.regionserver.minibatch.size";
            +241  public static final int 
            DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2;
             242
            -243  final AtomicBoolean closed = new 
            AtomicBoolean(false);
            -244
            -245  /* Closing can take some time; use the 
            closing flag if there is stuff we don't
            -246   * want to do while in closing state; 
            e.g. like offer this region up to the
            -247   * master as a region to close if the 
            carrying regionserver is overloaded.
            -248   * Once set, it is never cleared.
            -249   */
            -250  final AtomicBoolean closing = new 
            AtomicBoolean(false);
            -251
            -252  /**
            -253   * The max sequence id of flushed data 
            on this region. There is no edit in memory that is
            -254   * less that this sequence id.
            -255   */
            -256  private volatile long maxFlushedSeqId = 
            HConstants.NO_SEQNUM;
            -257
            -258  /**
            -259   * Record the sequence id of last flush 
            operation. Can be in advance of
            -260   * {@link #maxFlushedSeqId} when 
            flushing a single column family. In this case,
            -261   * {@link #maxFlushedSeqId} will be 
            older than the oldest edit in memory.
            -262   */
            -263  private volatile long lastFlushOpSeqId 
            = HConstants.NO_SEQNUM;
            -264
            -265  /**
            -266   * The sequence id of the last replayed 
            open region event from the primary region. This is used
            -267   * to skip entries before this due to 
            the possibility of replay edits coming out of order from
            -268   * replication.
            -269   */
            -270  protected volatile long 
            lastReplayedOpenRegionSeqId = -1L;
            -271  protected volatile long 
            lastReplayedCompactionSeqId = -1L;
            -272
            -273  
            //
            -274  // Members
            -275  
            //
            -276
            -277  // map from a locked row to the context 
            for that lock including:
            -278  // - CountDownLatch for threads waiting 
            on that row
            -279  // - the thread that owns the lock 
            (allow reentrancy)
            -280  // - reference count of (reentrant) 
            locks held by the thread
            -281  // - the row itself
            -282  private final 
            ConcurrentHashMapHashedBytes, RowLockContext lockedRows =
            -283  new ConcurrentHashMap();
            -284
            -285  protected final Mapbyte[], 
            HStore stores =
            -286  new 
            ConcurrentSkipListMap(Bytes.BYTES_RAWCOMPARATOR);
            +243  public static final String 
            WAL_HSYNC_CONF_KEY = "hbase.wal.hsync";
            +244  public static final boolean 
            DEFAULT_WAL_HSYNC = false;
            +245
            +246  final AtomicBoolean closed = new 
            AtomicBoolean(false);
            +247
            +248  /* Closing can take some time; use the 
            closing flag if there is stuff we don't
            +249   * want to do while in closing state; 
            e.g. like offer this region up to the
            +250   * master as a region to close if the 
            carrying regionserver is overloaded.
            +251   * Once set, it is never cleared.
            +252   */
            +253  final AtomicBoolean closing = new 
            AtomicBoolean(false);
            +254
            +255  /**
            +256   * The max sequence id of flushed data 
            on this region. There is no edit in memory that is
            +257   * less that this sequence id.
            +258   */
            +259  private volatile long maxFlushedSeqId = 
            HConstants.NO_SEQNUM;
            +260
            +261  /**
            +262   * Record the sequence id of last flush 
            operation. Can be in advance of
            +263   * {@link #maxFlushedSeqId} when 
            flushing a single column family. In this case,
            +264   * {@link #maxFlushedSeqId} will be 
            older than the oldest edit in memory.
            +265   */
            +266  

            [26/51] [partial] hbase-site git commit: Published site at 00095a2ef9442e3fd86c04876c9d91f2f8b23ad8.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.html
            index f47d627..c3d225c 100644
            --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.html
            +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.html
            @@ -117,219 +117,219 @@
             109   */
             110  public static boolean 
            archiveRegion(FileSystem fs, Path rootdir, Path tableDir, Path regionDir)
             111  throws IOException {
            -112if (LOG.isDebugEnabled()) {
            -113  LOG.debug("ARCHIVING " + 
            regionDir.toString());
            -114}
            -115
            -116// otherwise, we archive the files
            -117// make sure we can archive
            -118if (tableDir == null || regionDir == 
            null) {
            -119  LOG.error("No archive directory 
            could be found because tabledir (" + tableDir
            -120  + ") or regiondir (" + 
            regionDir + "was null. Deleting files instead.");
            -121  deleteRegionWithoutArchiving(fs, 
            regionDir);
            -122  // we should have archived, but 
            failed to. Doesn't matter if we deleted
            -123  // the archived files correctly or 
            not.
            -124  return false;
            -125}
            -126
            -127// make sure the regiondir lives 
            under the tabledir
            -128
            Preconditions.checkArgument(regionDir.toString().startsWith(tableDir.toString()));
            -129Path regionArchiveDir = 
            HFileArchiveUtil.getRegionArchiveDir(rootdir,
            -130FSUtils.getTableName(tableDir),
            -131regionDir.getName());
            -132
            -133FileStatusConverter getAsFile = new 
            FileStatusConverter(fs);
            -134// otherwise, we attempt to archive 
            the store files
            -135
            -136// build collection of just the store 
            directories to archive
            -137CollectionFile toArchive = 
            new ArrayList();
            -138final PathFilter dirFilter = new 
            FSUtils.DirFilter(fs);
            -139PathFilter nonHidden = new 
            PathFilter() {
            -140  @Override
            -141  public boolean accept(Path file) 
            {
            -142return dirFilter.accept(file) 
             !file.getName().toString().startsWith(".");
            -143  }
            -144};
            -145FileStatus[] storeDirs = 
            FSUtils.listStatus(fs, regionDir, nonHidden);
            -146// if there no files, we can just 
            delete the directory and return;
            -147if (storeDirs == null) {
            -148  LOG.debug("Region directory " + 
            regionDir + " empty.");
            -149  return 
            deleteRegionWithoutArchiving(fs, regionDir);
            -150}
            -151
            -152// convert the files in the region to 
            a File
            -153
            toArchive.addAll(Lists.transform(Arrays.asList(storeDirs), getAsFile));
            -154LOG.debug("Archiving " + 
            toArchive);
            -155ListFile failedArchive = 
            resolveAndArchive(fs, regionArchiveDir, toArchive,
            -156
            EnvironmentEdgeManager.currentTime());
            -157if (!failedArchive.isEmpty()) {
            -158  throw new 
            FailedArchiveException("Failed to archive/delete all the files for region:"
            -159  + regionDir.getName() + " into 
            " + regionArchiveDir
            -160  + ". Something is probably awry 
            on the filesystem.",
            -161  
            Collections2.transform(failedArchive, FUNC_FILE_TO_PATH));
            -162}
            -163// if that was successful, then we 
            delete the region
            -164return 
            deleteRegionWithoutArchiving(fs, regionDir);
            -165  }
            -166
            -167  /**
            -168   * Remove from the specified region the 
            store files of the specified column family,
            -169   * either by archiving them or outright 
            deletion
            -170   * @param fs the filesystem where the 
            store files live
            -171   * @param conf {@link Configuration} to 
            examine to determine the archive directory
            -172   * @param parent Parent region hosting 
            the store files
            -173   * @param tableDir {@link Path} to 
            where the table is being stored (for building the archive path)
            -174   * @param family the family hosting the 
            store files
            -175   * @throws IOException if the files 
            could not be correctly disposed.
            -176   */
            -177  public static void 
            archiveFamily(FileSystem fs, Configuration conf,
            -178  RegionInfo parent, Path tableDir, 
            byte[] family) throws IOException {
            -179Path familyDir = new Path(tableDir, 
            new Path(parent.getEncodedName(), Bytes.toString(family)));
            -180archiveFamilyByFamilyDir(fs, conf, 
            parent, familyDir, family);
            -181  }
            -182
            -183  /**
            -184   * Removes from the specified region 
            the store files of the specified column family,
            -185   * either by archiving them or outright 
            deletion
            -186   * @param fs the filesystem where the 
            store files live
            -187   * @param conf {@link Configuration} to 
            examine to determine the archive directory
            -188   * @param parent Parent region hosting 
            the store files
            -189   * @param familyDir {@link Path} to 
            where the family is being stored
            -190   * @param family the family hosting the 
            store files
            -191   * @throws IOException if the files 
            could not be correctly disposed.
            -192   */
            -193  public static void 
            archiveFamilyByFamilyDir(FileSystem fs, 

            [26/51] [partial] hbase-site git commit: Published site at 22f4def942f8a3367d0ca6598317e9b9a7d0cfcd.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.html 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.html
            index bf375f2..ad31b71 100644
            --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.html
            +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.html
            @@ -115,7 +115,7 @@ var activeTableTab = "activeTableTab";
             
             
             @InterfaceAudience.Private
            -public class RegionCoprocessorHost
            +public class RegionCoprocessorHost
             extends CoprocessorHostRegionCoprocessor,RegionCoprocessorEnvironment
             Implements the coprocessor environment and runtime support 
            for coprocessors
              loaded within a Region.
            @@ -796,7 +796,7 @@ extends 
             
             LOG
            -private static finalorg.slf4j.Logger LOG
            +private static finalorg.slf4j.Logger LOG
             
             
             
            @@ -805,7 +805,7 @@ extends 
             
             SHARED_DATA_MAP
            -private static 
            finalorg.apache.commons.collections4.map.ReferenceMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
             title="class or interface in java.util.concurrent">ConcurrentMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object SHARED_DATA_MAP
            +private static 
            finalorg.apache.commons.collections4.map.ReferenceMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
             title="class or interface in java.util.concurrent">ConcurrentMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object SHARED_DATA_MAP
             
             
             
            @@ -814,7 +814,7 @@ extends 
             
             hasCustomPostScannerFilterRow
            -private finalboolean hasCustomPostScannerFilterRow
            +private finalboolean hasCustomPostScannerFilterRow
             
             
             
            @@ -823,7 +823,7 @@ extends 
             
             rsServices
            -RegionServerServices rsServices
            +RegionServerServices rsServices
             The region server services
             
             
            @@ -833,7 +833,7 @@ extends 
             
             region
            -HRegion region
            +HRegion region
             The region
             
             
            @@ -843,7 +843,7 @@ extends 
             
             regionObserverGetter
            -privateCoprocessorHost.ObserverGetterRegionCoprocessor,RegionObserver 
            regionObserverGetter
            +privateCoprocessorHost.ObserverGetterRegionCoprocessor,RegionObserver 
            regionObserverGetter
             
             
             
            @@ -852,7 +852,7 @@ extends 
             
             endpointObserverGetter
            -privateCoprocessorHost.ObserverGetterRegionCoprocessor,EndpointObserver endpointObserverGetter
            +privateCoprocessorHost.ObserverGetterRegionCoprocessor,EndpointObserver endpointObserverGetter
             
             
             
            @@ -869,7 +869,7 @@ extends 
             
             RegionCoprocessorHost
            -publicRegionCoprocessorHost(HRegionregion,
            +publicRegionCoprocessorHost(HRegionregion,
              RegionServerServicesrsServices,
              
            org.apache.hadoop.conf.Configurationconf)
             Constructor
            @@ -895,7 +895,7 @@ extends 
             
             getTableCoprocessorAttrsFromSchema
            -statichttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionCoprocessorHost.TableCoprocessorAttributegetTableCoprocessorAttrsFromSchema(org.apache.hadoop.conf.Configurationconf,
            +statichttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionCoprocessorHost.TableCoprocessorAttributegetTableCoprocessorAttrsFromSchema(org.apache.hadoop.conf.Configurationconf,
            
             TableDescriptorhtd)
             
             
            @@ -905,7 +905,7 @@ extends 
             
             testTableCoprocessorAttrs
            -public staticvoidtestTableCoprocessorAttrs(org.apache.hadoop.conf.Configurationconf,
            +public staticvoidtestTableCoprocessorAttrs(org.apache.hadoop.conf.Configurationconf,
              TableDescriptorhtd)
               throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
             title="class or interface in java.io">IOException
             Sanity check the table coprocessor attributes of the 
            supplied schema. Will
            @@ -925,7 +925,7 @@ extends 
             
             loadTableCoprocessors
            -voidloadTableCoprocessors(org.apache.hadoop.conf.Configurationconf)
            

            [26/51] [partial] hbase-site git commit: Published site at 31da4d0bce69b3a47066a5df675756087ce4dc60.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/MoveRegionProcedure.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/MoveRegionProcedure.html 
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/MoveRegionProcedure.html
            index 5d49f63..561fb67 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/master/assignment/MoveRegionProcedure.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/master/assignment/MoveRegionProcedure.html
            @@ -134,12 +134,15 @@ var activeTableTab = "activeTableTab";
             
             
             @InterfaceAudience.Private
            -public class MoveRegionProcedure
            +public class MoveRegionProcedure
             extends AbstractStateMachineRegionProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MoveRegionState
             Procedure that implements a RegionPlan.
              It first runs an unassign subprocedure followed
              by an assign subprocedure. It takes a lock on the region being moved.
            - It holds the lock for the life of the procedure.
            + It holds the lock for the life of the procedure.
            +
            + Throws exception on construction if determines context hostile to move 
            (cluster going
            + down or master is shutting down or table is disabled).
             
             
             
            @@ -321,7 +324,7 @@ extends AbstractStateMachineTableProcedure
            -getRegionDir,
             getUser,
             releaseSyncLatch,
             setUser
            +getRegionDir,
             getUser,
             preflightChecks,
             releaseSyncLatch,
             setUser
             
             
             
            @@ -364,7 +367,7 @@ extends 
             
             LOG
            -private static finalorg.slf4j.Logger LOG
            +private static finalorg.slf4j.Logger LOG
             
             
             
            @@ -373,7 +376,7 @@ extends 
             
             plan
            -privateRegionPlan plan
            +privateRegionPlan plan
             
             
             
            @@ -390,7 +393,7 @@ extends 
             
             MoveRegionProcedure
            -publicMoveRegionProcedure()
            +publicMoveRegionProcedure()
             
             
             
            @@ -399,8 +402,15 @@ extends 
             
             MoveRegionProcedure
            -publicMoveRegionProcedure(MasterProcedureEnvenv,
            -   RegionPlanplan)
            +publicMoveRegionProcedure(MasterProcedureEnvenv,
            +   RegionPlanplan)
            +throws HBaseIOException
            +
            +Throws:
            +https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
             title="class or interface in java.io">IOException - If the cluster 
            is offline or master is stopping or if table is disabled
            +   or non-existent.
            +HBaseIOException
            +
             
             
             
            @@ -417,7 +427,7 @@ extends 
             
             executeFromState
            -protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv,
            +protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv,
               
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MoveRegionStatestate)
            throws https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
             title="class or interface in java.lang">InterruptedException
             Description copied from 
            class:StateMachineProcedure
            @@ -440,7 +450,7 @@ extends 
             
             rollbackState
            -protectedvoidrollbackState(MasterProcedureEnvenv,
            +protectedvoidrollbackState(MasterProcedureEnvenv,
              
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MoveRegionStatestate)
               throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
             title="class or interface in java.io">IOException
             Description copied from 
            class:StateMachineProcedure
            @@ -460,7 +470,7 @@ extends 
             
             abort
            -publicbooleanabort(MasterProcedureEnvenv)
            +publicbooleanabort(MasterProcedureEnvenv)
             Description copied from 
            class:Procedure
             The abort() call is asynchronous and each procedure must 
            decide how to deal
              with it, if they want to be abortable. The simplest implementation
            @@ -483,7 +493,7 @@ extends 
             
             toStringClassDetails
            -publicvoidtoStringClassDetails(https://docs.oracle.com/javase/8/docs/api/java/lang/StringBuilder.html?is-external=true;
             title="class or interface in java.lang">StringBuildersb)
            +publicvoidtoStringClassDetails(https://docs.oracle.com/javase/8/docs/api/java/lang/StringBuilder.html?is-external=true;
             title="class or interface in java.lang">StringBuildersb)
             Description copied from 
            class:Procedure
             Extend the toString() information with the procedure details
              e.g. className and parameters
            @@ -501,7 +511,7 @@ extends 
             
             getInitialState
            -protectedorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MoveRegionStategetInitialState()
            +protectedorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MoveRegionStategetInitialState()
             Description copied from 
            class:StateMachineProcedure
             Return the initial state object that will be used for the 
            first call to executeFromState().
             
            @@ -518,7 +528,7 @@ extends 
             
             getStateId
            -protectedintgetStateId(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MoveRegionStatestate)
            

            [26/51] [partial] hbase-site git commit: Published site at 6b77786dfc46d25ac5bb5f1c8a4a9eb47b52a604.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html
            index 6d14fc8..4b9fd29 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/CompactType.html
            @@ -127,30 +127,28 @@ the order they are declared.
             
             
             
            -private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            -RawAsyncHBaseAdmin.compact(TableNametableName,
            +private void
            +HBaseAdmin.compact(TableNametableName,
            byte[]columnFamily,
            booleanmajor,
            CompactTypecompactType)
            -Compact column family of a table, Asynchronous operation 
            even if CompletableFuture.get()
            +Compact a table.
             
             
             
            -private void
            -HBaseAdmin.compact(TableNametableName,
            +private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            +RawAsyncHBaseAdmin.compact(TableNametableName,
            byte[]columnFamily,
            booleanmajor,
            CompactTypecompactType)
            -Compact a table.
            +Compact column family of a table, Asynchronous operation 
            even if CompletableFuture.get()
             
             
             
             https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            -AsyncAdmin.compact(TableNametableName,
            +AsyncHBaseAdmin.compact(TableNametableName,
            byte[]columnFamily,
            -   CompactTypecompactType)
            -Compact a column family within a table.
            -
            +   CompactTypecompactType)
             
             
             void
            @@ -161,14 +159,16 @@ the order they are declared.
             
             
             
            -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            -RawAsyncHBaseAdmin.compact(TableNametableName,
            +void
            +HBaseAdmin.compact(TableNametableName,
            byte[]columnFamily,
            -   CompactTypecompactType)
            +   CompactTypecompactType)
            +Compact a column family within a table.
            +
             
             
            -void
            -HBaseAdmin.compact(TableNametableName,
            +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            +AsyncAdmin.compact(TableNametableName,
            byte[]columnFamily,
            CompactTypecompactType)
             Compact a column family within a table.
            @@ -176,16 +176,14 @@ the order they are declared.
             
             
             https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            -AsyncHBaseAdmin.compact(TableNametableName,
            +RawAsyncHBaseAdmin.compact(TableNametableName,
            byte[]columnFamily,
            CompactTypecompactType)
             
             
             https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            -AsyncAdmin.compact(TableNametableName,
            -   CompactTypecompactType)
            -Compact a table.
            -
            +AsyncHBaseAdmin.compact(TableNametableName,
            +   CompactTypecompactType)
             
             
             void
            @@ -195,28 +193,28 @@ the order they are declared.
             
             
             
            -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            -RawAsyncHBaseAdmin.compact(TableNametableName,
            -   CompactTypecompactType)
            -
            -
             void
             HBaseAdmin.compact(TableNametableName,
            CompactTypecompactType)
             Compact a table.
             
             
            +
            

            [26/51] [partial] hbase-site git commit: Published site at 1384da71375427b522b09f06862bb5d629cef52f.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            index 1d50582..7eb7661 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
            @@ -104,14 +104,14 @@
             
             
             void
            -HMaster.checkTableModifiable(TableNametableName)
            -
            -
            -void
             MasterServices.checkTableModifiable(TableNametableName)
             Check table is modifiable; i.e.
             
             
            +
            +void
            +HMaster.checkTableModifiable(TableNametableName)
            +
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            index 6194cbc..25d874c 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
            @@ -170,14 +170,14 @@
             
             
             void
            -HMaster.checkTableModifiable(TableNametableName)
            -
            -
            -void
             MasterServices.checkTableModifiable(TableNametableName)
             Check table is modifiable; i.e.
             
             
            +
            +void
            +HMaster.checkTableModifiable(TableNametableName)
            +
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html 
            b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            index 067aa54..22b3482 100644
            --- a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
            @@ -243,6 +243,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTag
            +PrivateCellUtil.getTags(Cellcell)
            +
            +
            +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTag
             CellUtil.getTags(Cellcell)
             Deprecated.
             As of release 2.0.0, this 
            will be removed in HBase 3.0.0.
            @@ -250,10 +254,6 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -
            -static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTag
            -PrivateCellUtil.getTags(Cellcell)
            -
             
             static https://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true;
             title="class or interface in java.util">IteratorTag
             CellUtil.tagsIterator(byte[]tags,
            @@ -395,6 +395,11 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             static Cell
            +PrivateCellUtil.createCell(Cellcell,
            +  https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
            +
            +
            +static Cell
             CellUtil.createCell(Cellcell,
               https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
             Deprecated.
            @@ -402,11 +407,6 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -
            -static Cell
            -PrivateCellUtil.createCell(Cellcell,
            -  https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
            -
             
             static byte[]
             TagUtil.fromList(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
            @@ -415,16 +415,16 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -RawCellBuilder
            -RawCellBuilder.setTags(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
            +ExtendedCellBuilder
            +ExtendedCellBuilderImpl.setTags(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
             
             
            -ExtendedCellBuilder
            -ExtendedCellBuilder.setTags(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
            +RawCellBuilder
            +RawCellBuilder.setTags(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTagtags)
             
             
             ExtendedCellBuilder
            

            [26/51] [partial] hbase-site git commit: Published site at b7b86839250bf9b295ebc1948826f43a88736d6c.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/org/apache/hadoop/hbase/replication/regionserver/WALEntryBatch.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/WALEntryBatch.html
             
            b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/WALEntryBatch.html
            new file mode 100644
            index 000..b8dc0b1
            --- /dev/null
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/WALEntryBatch.html
            @@ -0,0 +1,610 @@
            +http://www.w3.org/TR/html4/loose.dtd;>
            +
            +
            +
            +
            +
            +WALEntryBatch (Apache HBase 3.0.0-SNAPSHOT API)
            +
            +
            +
            +
            +
            +var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10};
            +var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
            +var altColor = "altColor";
            +var rowColor = "rowColor";
            +var tableTab = "tableTab";
            +var activeTableTab = "activeTableTab";
            +
            +
            +JavaScript is disabled on your browser.
            +
            +
            +
            +
            +
            +Skip navigation links
            +
            +
            +
            +
            +Overview
            +Package
            +Class
            +Use
            +Tree
            +Deprecated
            +Index
            +Help
            +
            +
            +
            +
            +PrevClass
            +NextClass
            +
            +
            +Frames
            +NoFrames
            +
            +
            +AllClasses
            +
            +
            +
            +
            +
            +
            +
            +Summary:
            +Nested|
            +Field|
            +Constr|
            +Method
            +
            +
            +Detail:
            +Field|
            +Constr|
            +Method
            +
            +
            +
            +
            +
            +
            +
            +
            +org.apache.hadoop.hbase.replication.regionserver
            +Class WALEntryBatch
            +
            +
            +
            +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">java.lang.Object
            +
            +
            +org.apache.hadoop.hbase.replication.regionserver.WALEntryBatch
            +
            +
            +
            +
            +
            +
            +
            +
            +@InterfaceAudience.Private
            +class WALEntryBatch
            +extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
            +Holds a batch of WAL entries to replicate, along with some 
            statistics
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +Field Summary
            +
            +Fields
            +
            +Modifier and Type
            +Field and Description
            +
            +
            +private long
            +heapSize
            +
            +
            +private https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
             title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
             title="class or interface in java.lang">Long
            +lastSeqIds
            +
            +
            +private org.apache.hadoop.fs.Path
            +lastWalPath
            +
            +
            +private long
            +lastWalPosition
            +
            +
            +private int
            +nbHFiles
            +
            +
            +private int
            +nbRowKeys
            +
            +
            +private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListWAL.Entry
            +walEntries
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +Constructor Summary
            +
            +Constructors
            +
            +Constructor and Description
            +
            +
            +WALEntryBatch(intmaxNbEntries,
            + org.apache.hadoop.fs.PathlastWalPath)
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +Method Summary
            +
            +All MethodsInstance MethodsConcrete Methods
            +
            +Modifier and Type
            +Method and Description
            +
            +
            +void
            +addEntry(WAL.Entryentry)
            +
            +
            +long
            +getHeapSize()
            +
            +
            +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
             title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
             title="class or interface in java.lang">Long
            +getLastSeqIds()
            +
            +
            +org.apache.hadoop.fs.Path
            +getLastWalPath()
            +
            +
            +long
            +getLastWalPosition()
            +
            +
            +int
            +getNbEntries()
            +
            +
            +int
            +getNbHFiles()
            +
            +
            +int
            +getNbOperations()
            +
            +
            +int
            +getNbRowKeys()
            +
            +
            +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListWAL.Entry
            +getWalEntries()
            +
            +
            +void
            +incrementHeapSize(longincrement)
            +
            +
            +void
            +incrementNbHFiles(intincrement)
            +
            +
            +void
            +incrementNbRowKeys(intincrement)
            +
            +
            +void
            +setLastSeqId(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringregion,
            +longsequenceId)
            +
            +
            +void
            +setLastWalPosition(longlastWalPosition)
            +
            +
            +
            +
            +
            +
            +Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
            +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
             title="class or interface in java.lang">clone, 

            [26/51] [partial] hbase-site git commit: Published site at 1d25b60831b8cc8f7ad5fd366f1867de5c20d2f3.

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/class-use/TableExistsException.html
            --
            diff --git 
            a/apidocs/org/apache/hadoop/hbase/class-use/TableExistsException.html 
            b/apidocs/org/apache/hadoop/hbase/class-use/TableExistsException.html
            index d8bdef1..611fa84 100644
            --- a/apidocs/org/apache/hadoop/hbase/class-use/TableExistsException.html
            +++ b/apidocs/org/apache/hadoop/hbase/class-use/TableExistsException.html
            @@ -113,22 +113,22 @@
             
             
             void
            -Admin.cloneSnapshot(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringsnapshotName,
            +Admin.cloneSnapshot(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringsnapshotName,
              TableNametableName)
             Create a new table by cloning the snapshot content.
             
             
             
             void
            -Admin.cloneSnapshot(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringsnapshotName,
            +Admin.cloneSnapshot(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringsnapshotName,
              TableNametableName,
              booleanrestoreAcl)
             Create a new table by cloning the snapshot content.
             
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Future.html?is-external=true;
             title="class or interface in java.util.concurrent">Futurehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            -Admin.cloneSnapshotAsync(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringsnapshotName,
            +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Future.html?is-external=true;
             title="class or interface in java.util.concurrent">Futurehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            +Admin.cloneSnapshotAsync(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringsnapshotName,
               TableNametableName)
             Create a new table by cloning the snapshot content, but 
            does not block
              and wait for it to be completely cloned.
            
            
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html 
            b/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
            index b99a6b3..f82f773 100644
            --- a/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
            +++ b/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
            @@ -197,8 +197,8 @@
             
             java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
             title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
             title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
             title="class or interface in java.io">Serializable)
             
            -org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.LocalityType
             org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type
            +org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.LocalityType
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html 
            b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
            index 81c4c95..becce8e 100644
            --- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
            +++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
            @@ -346,11 +346,11 @@
             
             java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
             title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
             title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
             title="class or interface in java.io">Serializable)
             
            -org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective
            -org.apache.hadoop.hbase.master.RegionState.State
            -org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus
             org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage
            +org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus
             org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode
            +org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective
            +org.apache.hadoop.hbase.master.RegionState.State
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/org/apache/hadoop/hbase/package-tree.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/package-tree.html 
            b/devapidocs/org/apache/hadoop/hbase/package-tree.html
            index 47d1eb6..e6678cc 100644
            --- a/devapidocs/org/apache/hadoop/hbase/package-tree.html
            +++ b/devapidocs/org/apache/hadoop/hbase/package-tree.html
            @@ -320,13 +320,13 @@
             org.apache.hadoop.hbase.UnknownScannerException
             
             
            +org.apache.hadoop.hbase.MasterNotRunningException
             org.apache.hadoop.hbase.PleaseHoldException
             org.apache.hadoop.hbase.RegionException
             org.apache.hadoop.hbase.ServiceNotRunningException
             org.apache.hadoop.hbase.TableInfoMissingException
             
             
            -org.apache.hadoop.hbase.MasterNotRunningException
             org.apache.hadoop.hbase.NotServingRegionException
             org.apache.hadoop.hbase.RegionTooBusyException
             org.apache.hadoop.hbase.RetryImmediatelyException
            @@ -440,20 +440,20 @@
             
             java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
             title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
             title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
             title="class or interface in java.io">Serializable)
             
            -org.apache.hadoop.hbase.HConstants.OperationStatusCode
            -org.apache.hadoop.hbase.CompareOperator
            +org.apache.hadoop.hbase.MetaTableAccessor.QueryType
             org.apache.hadoop.hbase.MemoryCompactionPolicy
            -org.apache.hadoop.hbase.Cell.Type
             org.apache.hadoop.hbase.Coprocessor.State
            -org.apache.hadoop.hbase.Size.Unit
            +org.apache.hadoop.hbase.CompareOperator
             org.apache.hadoop.hbase.ProcedureState
            -org.apache.hadoop.hbase.ClusterMetrics.Option
            -org.apache.hadoop.hbase.HealthChecker.HealthCheckerExitStatus
            -org.apache.hadoop.hbase.KeepDeletedCells
            -org.apache.hadoop.hbase.MetaTableAccessor.QueryType
            +org.apache.hadoop.hbase.HConstants.OperationStatusCode
            +org.apache.hadoop.hbase.KeyValue.Type
             org.apache.hadoop.hbase.CellBuilderType
            +org.apache.hadoop.hbase.HealthChecker.HealthCheckerExitStatus
             

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteCell.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteCell.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteCell.html
            index d143ef8..4583895 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteCell.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteCell.html
            @@ -258,7 +258,7 @@
             250
             251@Override
             252public long heapSize() {
            -253  long sum = HEAP_SIZE_OVERHEAD + 
            estimatedHeapSizeOf(cell);
            +253  long sum = HEAP_SIZE_OVERHEAD + 
            estimatedSizeOfCell(cell);
             254  if (this.tags != null) {
             255sum += 
            ClassSize.sizeOf(this.tags);
             256  }
            @@ -454,7 +454,7 @@
             446
             447@Override
             448public long heapSize() {
            -449  long sum = HEAP_SIZE_OVERHEAD + 
            estimatedHeapSizeOf(cell);
            +449  long sum = HEAP_SIZE_OVERHEAD + 
            estimatedSizeOfCell(cell);
             450  // this.tags is on heap byte[]
             451  if (this.tags != null) {
             452sum += 
            ClassSize.sizeOf(this.tags);
            @@ -2791,192 +2791,193 @@
             2783   * {@link HeapSize} we call {@link 
            HeapSize#heapSize()} so cell can give a correct value. In other
             2784   * cases we just consider the bytes 
            occupied by the cell components ie. row, CF, qualifier,
             2785   * timestamp, type, value and tags.
            -2786   * @param cell
            -2787   * @return estimate of the heap 
            space
            -2788   */
            -2789  public static long 
            estimatedHeapSizeOf(final Cell cell) {
            -2790if (cell instanceof HeapSize) {
            -2791  return ((HeapSize) 
            cell).heapSize();
            -2792}
            -2793// TODO: Add sizing of references 
            that hold the row, family, etc., arrays.
            -2794return 
            estimatedSerializedSizeOf(cell);
            -2795  }
            -2796
            -2797  /**
            -2798   * This method exists just to 
            encapsulate how we serialize keys. To be replaced by a factory that
            -2799   * we query to figure what the Cell 
            implementation is and then, what serialization engine to use
            -2800   * and further, how to serialize the 
            key for inclusion in hfile index. TODO.
            -2801   * @param cell
            -2802   * @return The key portion of the Cell 
            serialized in the old-school KeyValue way or null if passed
            -2803   * a null 
            codecell/code
            -2804   */
            -2805  public static byte[] 
            getCellKeySerializedAsKeyValueKey(final Cell cell) {
            -2806if (cell == null) return null;
            -2807byte[] b = new 
            byte[KeyValueUtil.keyLength(cell)];
            -2808KeyValueUtil.appendKeyTo(cell, b, 
            0);
            -2809return b;
            -2810  }
            -2811
            -2812  /**
            -2813   * Create a Cell that is smaller than 
            all other possible Cells for the given Cell's row.
            -2814   * @param cell
            -2815   * @return First possible Cell on 
            passed Cell's row.
            -2816   */
            -2817  public static Cell 
            createFirstOnRow(final Cell cell) {
            -2818if (cell instanceof 
            ByteBufferExtendedCell) {
            -2819  return new 
            FirstOnRowByteBufferExtendedCell(
            -2820  ((ByteBufferExtendedCell) 
            cell).getRowByteBuffer(),
            -2821  ((ByteBufferExtendedCell) 
            cell).getRowPosition(), cell.getRowLength());
            -2822}
            -2823return new 
            FirstOnRowCell(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength());
            -2824  }
            -2825
            -2826  public static Cell 
            createFirstOnRow(final byte[] row, int roffset, short rlength) {
            -2827return new FirstOnRowCell(row, 
            roffset, rlength);
            -2828  }
            -2829
            -2830  public static Cell 
            createFirstOnRow(final byte[] row, final byte[] family, final byte[] col) {
            -2831return createFirstOnRow(row, 0, 
            (short) row.length, family, 0, (byte) family.length, col, 0,
            -2832col.length);
            -2833  }
            -2834
            -2835  public static Cell 
            createFirstOnRow(final byte[] row, int roffset, short rlength,
            -2836  final byte[] family, int foffset, 
            byte flength, final byte[] col, int coffset, int clength) {
            -2837return new FirstOnRowColCell(row, 
            roffset, rlength, family, foffset, flength, col, coffset,
            -2838clength);
            -2839  }
            -2840
            -2841  public static Cell 
            createFirstOnRow(final byte[] row) {
            -2842return createFirstOnRow(row, 0, 
            (short) row.length);
            -2843  }
            -2844
            -2845  public static Cell 
            createFirstOnRowFamily(Cell cell, byte[] fArray, int foff, int flen) {
            -2846if (cell instanceof 
            ByteBufferExtendedCell) {
            -2847  return new 
            FirstOnRowColByteBufferExtendedCell(
            -2848  ((ByteBufferExtendedCell) 
            cell).getRowByteBuffer(),
            -2849  ((ByteBufferExtendedCell) 
            cell).getRowPosition(), cell.getRowLength(),
            -2850  ByteBuffer.wrap(fArray), foff, 
            (byte) flen, HConstants.EMPTY_BYTE_BUFFER, 0, 0);
            -2851}
            -2852return new 
            FirstOnRowColCell(cell.getRowArray(), cell.getRowOffset(), 
            cell.getRowLength(),
            -2853fArray, foff, (byte) flen, 
            HConstants.EMPTY_BYTE_ARRAY, 0, 0);
            -2854  }
            -2855
            -2856  public static Cell 
            createFirstOnRowCol(final Cell cell) {
            -2857if (cell instanceof 
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.html
            index f433932..112afdf 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RemoteWithExtrasException.html
            @@ -27,103 +27,110 @@
             019
             020import java.io.IOException;
             021import java.lang.reflect.Constructor;
            -022
            -023import 
            org.apache.hadoop.conf.Configuration;
            -024import 
            org.apache.hadoop.hbase.DoNotRetryIOException;
            -025import 
            org.apache.hadoop.hbase.HBaseConfiguration;
            -026import 
            org.apache.yetus.audience.InterfaceAudience;
            -027import 
            org.apache.hadoop.hbase.protobuf.ProtobufUtil;
            -028import 
            org.apache.hadoop.hbase.util.DynamicClassLoader;
            -029import 
            org.apache.hadoop.ipc.RemoteException;
            -030
            -031/**
            -032 * A {@link RemoteException} with some 
            extra information.  If source exception
            -033 * was a {@link 
            org.apache.hadoop.hbase.DoNotRetryIOException}, 
            -034 * {@link #isDoNotRetry()} will return 
            true.
            -035 * pA {@link RemoteException} 
            hosts exceptions we got from the server.
            -036 */
            -037@SuppressWarnings("serial")
            -038@InterfaceAudience.Public
            -039@edu.umd.cs.findbugs.annotations.SuppressWarnings(
            -040value = 
            "DP_CREATE_CLASSLOADER_INSIDE_DO_PRIVILEGED", justification = "None. Address 
            sometime.")
            +022import java.security.AccessController;
            +023import java.security.PrivilegedAction;
            +024
            +025import 
            org.apache.hadoop.conf.Configuration;
            +026import 
            org.apache.hadoop.hbase.DoNotRetryIOException;
            +027import 
            org.apache.hadoop.hbase.HBaseConfiguration;
            +028import 
            org.apache.yetus.audience.InterfaceAudience;
            +029import 
            org.apache.hadoop.hbase.protobuf.ProtobufUtil;
            +030import 
            org.apache.hadoop.hbase.util.DynamicClassLoader;
            +031import 
            org.apache.hadoop.ipc.RemoteException;
            +032
            +033/**
            +034 * A {@link RemoteException} with some 
            extra information.  If source exception
            +035 * was a {@link 
            org.apache.hadoop.hbase.DoNotRetryIOException}, 
            +036 * {@link #isDoNotRetry()} will return 
            true.
            +037 * pA {@link RemoteException} 
            hosts exceptions we got from the server.
            +038 */
            +039@SuppressWarnings("serial")
            +040@InterfaceAudience.Public
             041public class RemoteWithExtrasException 
            extends RemoteException {
             042  private final String hostname;
             043  private final int port;
             044  private final boolean doNotRetry;
             045
            -046  private final static ClassLoader 
            CLASS_LOADER;
            -047
            -048  static {
            -049ClassLoader parent = 
            RemoteWithExtrasException.class.getClassLoader();
            -050Configuration conf = 
            HBaseConfiguration.create();
            -051CLASS_LOADER = new 
            DynamicClassLoader(conf, parent);
            -052  }
            -053
            -054  public RemoteWithExtrasException(String 
            className, String msg, final boolean doNotRetry) {
            -055this(className, msg, null, -1, 
            doNotRetry);
            -056  }
            -057
            -058  public RemoteWithExtrasException(String 
            className, String msg, final String hostname,
            -059  final int port, final boolean 
            doNotRetry) {
            -060super(className, msg);
            -061this.hostname = hostname;
            -062this.port = port;
            -063this.doNotRetry = doNotRetry;
            -064  }
            -065
            -066  @Override
            -067  public IOException 
            unwrapRemoteException() {
            -068Class? realClass;
            -069try {
            -070  // try to load a exception class 
            from where the HBase classes are loaded or from Dynamic
            -071  // classloader.
            -072  realClass = 
            Class.forName(getClassName(), false, CLASS_LOADER);
            -073} catch (ClassNotFoundException cnfe) 
            {
            -074  try {
            -075// cause could be a hadoop 
            exception, try to load from hadoop classpath
            -076realClass = 
            Class.forName(getClassName(), false, super.getClass().getClassLoader());
            -077  } catch (ClassNotFoundException e) 
            {
            -078return new 
            DoNotRetryIOException(
            -079"Unable to load exception 
            received from server:" + e.getMessage(), this);
            -080  }
            -081}
            -082try {
            -083  return 
            instantiateException(realClass.asSubclass(IOException.class));
            -084} catch (Exception e) {
            -085  return new DoNotRetryIOException(
            -086  "Unable to instantiate 
            exception received from server:" + e.getMessage(), this);
            -087}
            -088  }
            -089
            -090  private IOException 
            instantiateException(Class? extends IOException cls) throws Exception 
            {
            -091Constructor? extends 
            IOException cn = cls.getConstructor(String.class);
            -092cn.setAccessible(true);
            -093IOException ex = 
            cn.newInstance(this.getMessage());
            -094ex.initCause(this);
            -095return ex;
            -096  }
            -097
            -098  /**
            -099   * @return null if not set
            -100   */
            -101  public String getHostname() {
            -102return this.hostname;
            +046  /**
            +047   * Dynamic class loader to load 
            filter/comparators
            +048   */
            +049  private final 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html
            index 78d979d..81b1f23 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html
            @@ -292,7 +292,7 @@ service.
             
             
             private static HRegionLocation
            -MetaTableAccessor.getRegionLocation(Resultr,
            +AsyncMetaTableAccessor.getRegionLocation(Resultr,
              RegionInforegionInfo,
              intreplicaId)
             Returns the HRegionLocation parsed from the given meta row 
            Result
            @@ -301,7 +301,7 @@ service.
             
             
             private static HRegionLocation
            -AsyncMetaTableAccessor.getRegionLocation(Resultr,
            +MetaTableAccessor.getRegionLocation(Resultr,
              RegionInforegionInfo,
              intreplicaId)
             Returns the HRegionLocation parsed from the given meta row 
            Result
            @@ -309,14 +309,14 @@ service.
             
             
             
            -static RegionLocations
            -MetaTableAccessor.getRegionLocations(Resultr)
            +private static http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
             title="class or interface in java.util">OptionalRegionLocations
            +AsyncMetaTableAccessor.getRegionLocations(Resultr)
             Returns an HRegionLocationList extracted from the 
            result.
             
             
             
            -private static http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
             title="class or interface in java.util">OptionalRegionLocations
            -AsyncMetaTableAccessor.getRegionLocations(Resultr)
            +static RegionLocations
            +MetaTableAccessor.getRegionLocations(Resultr)
             Returns an HRegionLocationList extracted from the 
            result.
             
             
            @@ -326,42 +326,42 @@ service.
             
             
             private static long
            -MetaTableAccessor.getSeqNumDuringOpen(Resultr,
            +AsyncMetaTableAccessor.getSeqNumDuringOpen(Resultr,
            intreplicaId)
             The latest seqnum that the server writing to meta observed 
            when opening the region.
             
             
             
             private static long
            -AsyncMetaTableAccessor.getSeqNumDuringOpen(Resultr,
            +MetaTableAccessor.getSeqNumDuringOpen(Resultr,
            intreplicaId)
             The latest seqnum that the server writing to meta observed 
            when opening the region.
             
             
             
            -static ServerName
            -MetaTableAccessor.getServerName(Resultr,
            +private static http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
             title="class or interface in java.util">OptionalServerName
            +AsyncMetaTableAccessor.getServerName(Resultr,
              intreplicaId)
             Returns a ServerName from catalog table Result.
             
             
             
            -private static http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
             title="class or interface in java.util">OptionalServerName
            -AsyncMetaTableAccessor.getServerName(Resultr,
            +static ServerName
            +MetaTableAccessor.getServerName(Resultr,
              intreplicaId)
             Returns a ServerName from catalog table Result.
             
             
             
            +private static http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
             title="class or interface in java.util">OptionalTableState
            +AsyncMetaTableAccessor.getTableState(Resultr)
            +
            +
             static TableState
             MetaTableAccessor.getTableState(Resultr)
             Decode table state from META Result.
             
             
            -
            -private static http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
             title="class or interface in java.util">OptionalTableState
            -AsyncMetaTableAccessor.getTableState(Resultr)
            -
             
             void
             AsyncMetaTableAccessor.MetaTableScanResultConsumer.onNext(Result[]results,
            @@ -457,13 +457,13 @@ service.
             ClientScanner.cache
             
             
            -private http://docs.oracle.com/javase/8/docs/api/java/util/Deque.html?is-external=true;
             title="class or interface in java.util">DequeResult
            -BatchScanResultCache.partialResults
            -
            -
             private http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListResult
             CompleteScanResultCache.partialResults
             
            +
            +private http://docs.oracle.com/javase/8/docs/api/java/util/Deque.html?is-external=true;
             title="class or interface in java.util">DequeResult
            +BatchScanResultCache.partialResults
            +
             
             private http://docs.oracle.com/javase/8/docs/api/java/util/Queue.html?is-external=true;
             title="class or interface in java.util">QueueResult
             AsyncTableResultScanner.queue
            @@ -486,7 +486,7 @@ service.
             
             
             Result[]
            -BatchScanResultCache.addAndGet(Result[]results,
            +AllowPartialScanResultCache.addAndGet(Result[]results,
              booleanisHeartbeatMessage)
             
             
            @@ -496,20 +496,24 @@ service.
             
             
             Result[]
            -AllowPartialScanResultCache.addAndGet(Result[]results,
            +BatchScanResultCache.addAndGet(Result[]results,
              booleanisHeartbeatMessage)
             
             
             Result
            -HTable.append(Appendappend)
            -
            -
            -Result
             Table.append(Appendappend)
             Appends values to one or more columns within a single 
            row.
             
             
            +
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
            index 1fa011e..720740a 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
            @@ -849,23 +849,23 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             private Connection
            -RestoreTablesClient.conn
            +BackupAdminImpl.conn
             
             
            -protected Connection
            -TableBackupClient.conn
            -
            -
             (package private) Connection
             BackupCommands.Command.conn
             
            +
            +private Connection
            +RestoreTablesClient.conn
            +
             
             protected Connection
            -BackupManager.conn
            +TableBackupClient.conn
             
             
            -private Connection
            -BackupAdminImpl.conn
            +protected Connection
            +BackupManager.conn
             
             
             private Connection
            @@ -1179,13 +1179,13 @@ Input/OutputFormats, a table indexing MapReduce job, 
            and utility methods.
             
             
             
            -(package private) Connection
            -ConnectionImplementation.MasterServiceState.connection
            -
            -
             private Connection
             RegionServerCallable.connection
             
            +
            +(package private) Connection
            +ConnectionImplementation.MasterServiceState.connection
            +
             
             
             
            @@ -1230,20 +1230,20 @@ Input/OutputFormats, a table indexing MapReduce job, 
            and utility methods.
             
             
             
            -Connection
            -Admin.getConnection()
            -
            -
             (package private) Connection
             RegionAdminServiceCallable.getConnection()
             
            -
            +
             protected Connection
             HTable.getConnection()
             INTERNAL Used by unit tests and tools to do 
            low-level
              manipulations.
             
             
            +
            +Connection
            +Admin.getConnection()
            +
             
             Connection
             HBaseAdmin.getConnection()
            @@ -1557,11 +1557,11 @@ Input/OutputFormats, a table indexing MapReduce job, 
            and utility methods.
             
             
             private Connection
            -TableInputFormatBase.connection
            +HRegionPartitioner.connection
             
             
             private Connection
            -HRegionPartitioner.connection
            +TableInputFormatBase.connection
             
             
             
            @@ -1594,22 +1594,22 @@ Input/OutputFormats, a table indexing MapReduce job, 
            and utility methods.
             
             
             
            -private Connection
            -TableOutputFormat.TableRecordWriter.connection
            -
            -
             (package private) Connection
             MultiTableOutputFormat.MultiTableRecordWriter.connection
             
            +
            +private Connection
            +HRegionPartitioner.connection
            +
             
             private Connection
            -TableInputFormatBase.connection
            -The underlying Connection 
            of the table.
            -
            +TableOutputFormat.TableRecordWriter.connection
             
             
             private Connection
            -HRegionPartitioner.connection
            +TableInputFormatBase.connection
            +The underlying Connection 
            of the table.
            +
             
             
             (package private) Connection
            @@ -1694,15 +1694,15 @@ Input/OutputFormats, a table indexing MapReduce job, 
            and utility methods.
             
             
             private Connection
            -CatalogJanitor.connection
            +RegionPlacementMaintainer.connection
             
             
             private Connection
            -SnapshotOfRegionAssignmentFromMeta.connection
            +CatalogJanitor.connection
             
             
             private Connection
            -RegionPlacementMaintainer.connection
            +SnapshotOfRegionAssignmentFromMeta.connection
             
             
             
            @@ -1839,31 +1839,31 @@ Input/OutputFormats, a table indexing MapReduce job, 
            and utility methods.
             
             
             private Connection
            -TableQuotaSnapshotStore.conn
            +QuotaObserverChore.conn
             
             
             private Connection
            -SpaceQuotaRefresherChore.conn
            +QuotaObserverChore.TablesWithQuotas.conn
             
             
             private Connection
            -NamespaceQuotaSnapshotStore.conn
            +SnapshotQuotaObserverChore.conn
             
             
             private Connection
            -SnapshotQuotaObserverChore.conn
            +NamespaceQuotaSnapshotStore.conn
             
             
             private Connection
            -QuotaObserverChore.conn
            +TableQuotaSnapshotStore.conn
             
             
             private Connection
            -QuotaObserverChore.TablesWithQuotas.conn
            +TableSpaceQuotaSnapshotNotifier.conn
             
             
             private Connection
            -TableSpaceQuotaSnapshotNotifier.conn
            +SpaceQuotaRefresherChore.conn
             
             
             private Connection
            @@ -2197,11 +2197,11 @@ Input/OutputFormats, a table indexing MapReduce job, 
            and utility methods.
             
             
             Connection
            -RegionCoprocessorHost.RegionEnvironment.createConnection(org.apache.hadoop.conf.Configurationconf)
            +HRegionServer.createConnection(org.apache.hadoop.conf.Configurationconf)
             
             
             Connection
            -HRegionServer.createConnection(org.apache.hadoop.conf.Configurationconf)
            +RegionCoprocessorHost.RegionEnvironment.createConnection(org.apache.hadoop.conf.Configurationconf)
             
             
             Connection
            @@ -2209,11 +2209,11 @@ Input/OutputFormats, a table indexing MapReduce job, 
            and utility methods.
             
             
             Connection
            -RegionCoprocessorHost.RegionEnvironment.getConnection()
            +HRegionServer.getConnection()
             
             
             Connection
            -HRegionServer.getConnection()
            +RegionCoprocessorHost.RegionEnvironment.getConnection()
             
             
             
            @@ -2247,11 +2247,11 @@ Input/OutputFormats, a table indexing MapReduce job, 
            and utility methods.
             
             
             private Connection
            -HFileReplicator.connection
            +ReplicationSourceManager.connection
             
             
             private Connection
            -ReplicationSourceManager.connection
            +HFileReplicator.connection
             
             
             private 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            index aeaf9fe..232a8b4 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            @@ -495,7 +495,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             private static HRegionLocation
            -MetaTableAccessor.getRegionLocation(Resultr,
            +AsyncMetaTableAccessor.getRegionLocation(Resultr,
              RegionInforegionInfo,
              intreplicaId)
             Returns the HRegionLocation parsed from the given meta row 
            Result
            @@ -504,7 +504,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             private static HRegionLocation
            -AsyncMetaTableAccessor.getRegionLocation(Resultr,
            +MetaTableAccessor.getRegionLocation(Resultr,
              RegionInforegionInfo,
              intreplicaId)
             Returns the HRegionLocation parsed from the given meta row 
            Result
            @@ -944,9 +944,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -AsyncAdmin.getRegions(ServerNameserverName)
            -Get all the online regions on a region server.
            -
            +AsyncHBaseAdmin.getRegions(ServerNameserverName)
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            @@ -955,22 +953,22 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
            -
            -
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
             HBaseAdmin.getRegions(ServerNamesn)
             
            +
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            +AsyncAdmin.getRegions(ServerNameserverName)
            +Get all the online regions on a region server.
            +
            +
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -AsyncHBaseAdmin.getRegions(ServerNameserverName)
            +RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -AsyncAdmin.getRegions(TableNametableName)
            -Get the regions of a given table.
            -
            +AsyncHBaseAdmin.getRegions(TableNametableName)
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            @@ -979,16 +977,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -RawAsyncHBaseAdmin.getRegions(TableNametableName)
            -
            -
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
             HBaseAdmin.getRegions(TableNametableName)
             
            +
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncBufferedMutatorBuilder.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncBufferedMutatorBuilder.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncBufferedMutatorBuilder.html
            index 9f3035a..f2c8f10 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncBufferedMutatorBuilder.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncBufferedMutatorBuilder.html
            @@ -121,26 +121,26 @@
             
             
             AsyncBufferedMutatorBuilder
            +AsyncConnectionImpl.getBufferedMutatorBuilder(TableNametableName)
            +
            +
            +AsyncBufferedMutatorBuilder
             AsyncConnection.getBufferedMutatorBuilder(TableNametableName)
             Returns an AsyncBufferedMutatorBuilder 
            for creating AsyncBufferedMutator.
             
             
            -
            +
             AsyncBufferedMutatorBuilder
            -AsyncConnectionImpl.getBufferedMutatorBuilder(TableNametableName)
            +AsyncConnectionImpl.getBufferedMutatorBuilder(TableNametableName,
            + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
             title="class or interface in 
            java.util.concurrent">ExecutorServicepool)
             
            -
            +
             AsyncBufferedMutatorBuilder
             AsyncConnection.getBufferedMutatorBuilder(TableNametableName,
              http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
             title="class or interface in 
            java.util.concurrent">ExecutorServicepool)
             Returns an AsyncBufferedMutatorBuilder 
            for creating AsyncBufferedMutator.
             
             
            -
            -AsyncBufferedMutatorBuilder
            -AsyncConnectionImpl.getBufferedMutatorBuilder(TableNametableName,
            - http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
             title="class or interface in 
            java.util.concurrent">ExecutorServicepool)
            -
             
             AsyncBufferedMutatorBuilder
             AsyncBufferedMutatorBuilderImpl.setMaxAttempts(intmaxAttempts)
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
            index 024eca4..5ba2deb 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
            @@ -106,11 +106,11 @@
             
             
             private AsyncConnectionImpl
            -AsyncClientScanner.conn
            +RawAsyncTableImpl.conn
             
             
             private AsyncConnectionImpl
            -AsyncRpcRetryingCallerFactory.conn
            +AsyncBatchRpcRetryingCaller.conn
             
             
             private AsyncConnectionImpl
            @@ -118,19 +118,19 @@
             
             
             private AsyncConnectionImpl
            -RawAsyncTableImpl.conn
            +RegionCoprocessorRpcChannelImpl.conn
             
             
            -private AsyncConnectionImpl
            -RegionCoprocessorRpcChannelImpl.conn
            +protected AsyncConnectionImpl
            +AsyncRpcRetryingCaller.conn
             
             
             private AsyncConnectionImpl
            -AsyncBatchRpcRetryingCaller.conn
            +AsyncClientScanner.conn
             
             
            -protected AsyncConnectionImpl
            -AsyncRpcRetryingCaller.conn
            +private AsyncConnectionImpl
            +AsyncRpcRetryingCallerFactory.conn
             
             
             private AsyncConnectionImpl
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
            index d6b1759..e71ca45 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
            @@ -105,13 +105,13 @@
             
             
             
            -private AsyncMasterRequestRpcRetryingCaller.CallableT
            -AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder.callable
            -
            -
             private AsyncMasterRequestRpcRetryingCaller.CallableT
             AsyncMasterRequestRpcRetryingCaller.callable
             
            +
            +private AsyncMasterRequestRpcRetryingCaller.CallableT
            +AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder.callable
            +
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
            index f31564e..60fbcff 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
            +++ 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            index aeaf9fe..232a8b4 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
            @@ -495,7 +495,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             private static HRegionLocation
            -MetaTableAccessor.getRegionLocation(Resultr,
            +AsyncMetaTableAccessor.getRegionLocation(Resultr,
              RegionInforegionInfo,
              intreplicaId)
             Returns the HRegionLocation parsed from the given meta row 
            Result
            @@ -504,7 +504,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             private static HRegionLocation
            -AsyncMetaTableAccessor.getRegionLocation(Resultr,
            +MetaTableAccessor.getRegionLocation(Resultr,
              RegionInforegionInfo,
              intreplicaId)
             Returns the HRegionLocation parsed from the given meta row 
            Result
            @@ -944,9 +944,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -AsyncAdmin.getRegions(ServerNameserverName)
            -Get all the online regions on a region server.
            -
            +AsyncHBaseAdmin.getRegions(ServerNameserverName)
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            @@ -955,22 +953,22 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
            -
            -
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
             HBaseAdmin.getRegions(ServerNamesn)
             
            +
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            +AsyncAdmin.getRegions(ServerNameserverName)
            +Get all the online regions on a region server.
            +
            +
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -AsyncHBaseAdmin.getRegions(ServerNameserverName)
            +RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -AsyncAdmin.getRegions(TableNametableName)
            -Get the regions of a given table.
            -
            +AsyncHBaseAdmin.getRegions(TableNametableName)
             
             
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            @@ -979,16 +977,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
            utility methods.
             
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            -RawAsyncHBaseAdmin.getRegions(TableNametableName)
            -
            -
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
             HBaseAdmin.getRegions(TableNametableName)
             
            +
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
             title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListRegionInfo
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
            index 2ac1b78..90f52b0 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
            @@ -106,7 +106,7 @@
             
             
             private RegionLocateType
            -AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.locateType
            +AsyncSingleRequestRpcRetryingCaller.locateType
             
             
             RegionLocateType
            @@ -114,7 +114,7 @@
             
             
             private RegionLocateType
            -AsyncSingleRequestRpcRetryingCaller.locateType
            +AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.locateType
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
            index fbe0658..e062eb5 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
            @@ -230,13 +230,13 @@ service.
             
             
             private RegionLocator
            -TableInputFormatBase.regionLocator
            -The RegionLocator of the 
            table.
            -
            +HFileOutputFormat2.TableInfo.regionLocator
             
             
             private RegionLocator
            -HFileOutputFormat2.TableInfo.regionLocator
            +TableInputFormatBase.regionLocator
            +The RegionLocator of the 
            table.
            +
             
             
             
            @@ -248,15 +248,15 @@ service.
             
             
             
            +RegionLocator
            +HFileOutputFormat2.TableInfo.getRegionLocator()
            +
            +
             protected RegionLocator
             TableInputFormatBase.getRegionLocator()
             Allows subclasses to get the RegionLocator.
             
             
            -
            -RegionLocator
            -HFileOutputFormat2.TableInfo.getRegionLocator()
            -
             
             
             
            
            
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html
            index ad601c4..53e455f 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html
            @@ -1117,1183 +1117,1186 @@
             1109  @Nullable
             1110  public static TableState 
            getTableState(Connection conn, TableName tableName)
               throws IOException {
            -1112Table metaHTable = 
            getMetaHTable(conn);
            -1113Get get = new 
            Get(tableName.getName()).addColumn(getTableFamily(), getTableStateColumn());
            -1114long time = 
            EnvironmentEdgeManager.currentTime();
            -1115get.setTimeRange(0, time);
            -1116Result result =
            -1117metaHTable.get(get);
            -1118return getTableState(result);
            -1119  }
            -1120
            -1121  /**
            -1122   * Fetch table states from META 
            table
            -1123   * @param conn connection to use
            -1124   * @return map {tableName -gt; 
            state}
            -1125   * @throws IOException
            -1126   */
            -1127  public static MapTableName, 
            TableState getTableStates(Connection conn)
            -1128  throws IOException {
            -1129final MapTableName, 
            TableState states = new LinkedHashMap();
            -1130Visitor collector = new Visitor() 
            {
            -1131  @Override
            -1132  public boolean visit(Result r) 
            throws IOException {
            -1133TableState state = 
            getTableState(r);
            -1134if (state != null)
            -1135  
            states.put(state.getTableName(), state);
            -1136return true;
            -1137  }
            -1138};
            -1139fullScanTables(conn, collector);
            -1140return states;
            -1141  }
            -1142
            -1143  /**
            -1144   * Updates state in META
            -1145   * @param conn connection to use
            -1146   * @param tableName table to look 
            for
            -1147   * @throws IOException
            -1148   */
            -1149  public static void 
            updateTableState(Connection conn, TableName tableName,
            -1150  TableState.State actual) throws 
            IOException {
            -1151updateTableState(conn, new 
            TableState(tableName, actual));
            -1152  }
            -1153
            -1154  /**
            -1155   * Decode table state from META 
            Result.
            -1156   * Should contain cell from 
            HConstants.TABLE_FAMILY
            -1157   * @param r result
            -1158   * @return null if not found
            -1159   * @throws IOException
            -1160   */
            -1161  @Nullable
            -1162  public static TableState 
            getTableState(Result r)
            -1163  throws IOException {
            -1164Cell cell = 
            r.getColumnLatestCell(getTableFamily(), getTableStateColumn());
            -1165if (cell == null) return null;
            -1166try {
            -1167  return 
            TableState.parseFrom(TableName.valueOf(r.getRow()),
            -1168  
            Arrays.copyOfRange(cell.getValueArray(),
            -1169  cell.getValueOffset(), 
            cell.getValueOffset() + cell.getValueLength()));
            -1170} catch (DeserializationException e) 
            {
            -1171  throw new IOException(e);
            -1172}
            -1173
            -1174  }
            -1175
            -1176  /**
            -1177   * Implementations 'visit' a catalog 
            table row.
            -1178   */
            -1179  public interface Visitor {
            -1180/**
            -1181 * Visit the catalog table row.
            -1182 * @param r A row from catalog 
            table
            -1183 * @return True if we are to proceed 
            scanning the table, else false if
            -1184 * we are to stop now.
            -1185 */
            -1186boolean visit(final Result r) throws 
            IOException;
            -1187  }
            -1188
            -1189  /**
            -1190   * Implementations 'visit' a catalog 
            table row but with close() at the end.
            -1191   */
            -1192  public interface CloseableVisitor 
            extends Visitor, Closeable {
            -1193  }
            -1194
            -1195  /**
            -1196   * A {@link Visitor} that collects 
            content out of passed {@link Result}.
            -1197   */
            -1198  static abstract class 
            CollectingVisitorT implements Visitor {
            -1199final ListT results = new 
            ArrayList();
            -1200@Override
            -1201public boolean visit(Result r) 
            throws IOException {
            -1202  if (r ==  null || r.isEmpty()) 
            return true;
            -1203  add(r);
            -1204  return true;
            -1205}
            -1206
            -1207abstract void add(Result r);
            -1208
            -1209/**
            -1210 * @return Collected results; wait 
            till visits complete to collect all
            -1211 * possible results
            -1212 */
            -1213ListT getResults() {
            -1214  return this.results;
            -1215}
            -1216  }
            -1217
            -1218  /**
            -1219   * Collects all returned.
            -1220   */
            -1221  static class CollectAllVisitor extends 
            CollectingVisitorResult {
            -1222@Override
            -1223void add(Result r) {
            -1224  this.results.add(r);
            -1225}
            -1226  }
            -1227
            -1228  /**
            -1229   * A Visitor that skips offline 
            regions and split parents
            -1230   */
            -1231  public static abstract class 
            DefaultVisitorBase implements Visitor {
            -1232
            -1233public DefaultVisitorBase() {
            -1234  super();
            -1235}
            -1236
            -1237public abstract boolean 
            visitInternal(Result rowResult) throws IOException;
            -1238
            -1239@Override
            -1240public boolean visit(Result 
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/util/HBaseFsck.HbckInfo.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/util/HBaseFsck.HbckInfo.html 
            b/devapidocs/org/apache/hadoop/hbase/util/HBaseFsck.HbckInfo.html
            index 7f50b5d..7d756b9 100644
            --- a/devapidocs/org/apache/hadoop/hbase/util/HBaseFsck.HbckInfo.html
            +++ b/devapidocs/org/apache/hadoop/hbase/util/HBaseFsck.HbckInfo.html
            @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -public static class HBaseFsck.HbckInfo
            +public static class HBaseFsck.HbckInfo
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
             implements KeyRange
             Maintain information about a particular region.  It gathers 
            information
            @@ -305,7 +305,7 @@ implements 
             
             metaEntry
            -privateHBaseFsck.MetaEntry metaEntry
            +privateHBaseFsck.MetaEntry metaEntry
             
             
             
            @@ -314,7 +314,7 @@ implements 
             
             hdfsEntry
            -privateHBaseFsck.HdfsEntry hdfsEntry
            +privateHBaseFsck.HdfsEntry hdfsEntry
             
             
             
            @@ -323,7 +323,7 @@ implements 
             
             deployedEntries
            -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListHBaseFsck.OnlineEntry deployedEntries
            +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListHBaseFsck.OnlineEntry deployedEntries
             
             
             
            @@ -332,7 +332,7 @@ implements 
             
             deployedOn
            -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListServerName deployedOn
            +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListServerName deployedOn
             
             
             
            @@ -341,7 +341,7 @@ implements 
             
             skipChecks
            -privateboolean skipChecks
            +privateboolean skipChecks
             
             
             
            @@ -350,7 +350,7 @@ implements 
             
             isMerged
            -privateboolean isMerged
            +privateboolean isMerged
             
             
             
            @@ -359,7 +359,7 @@ implements 
             
             deployedReplicaId
            -privateint deployedReplicaId
            +privateint deployedReplicaId
             
             
             
            @@ -368,7 +368,7 @@ implements 
             
             primaryHRIForDeployedReplica
            -privateRegionInfo primaryHRIForDeployedReplica
            +privateRegionInfo primaryHRIForDeployedReplica
             
             
             
            @@ -385,7 +385,7 @@ implements 
             
             HbckInfo
            -HbckInfo(HBaseFsck.MetaEntrymetaEntry)
            +HbckInfo(HBaseFsck.MetaEntrymetaEntry)
             
             
             
            @@ -402,7 +402,7 @@ implements 
             
             getReplicaId
            -publicintgetReplicaId()
            +publicintgetReplicaId()
             
             
             
            @@ -411,7 +411,7 @@ implements 
             
             addServer
            -publicvoidaddServer(RegionInfohri,
            +publicvoidaddServer(RegionInfohri,
               ServerNameserver)
             
             
            @@ -421,7 +421,7 @@ implements 
             
             toString
            -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringtoString()
            +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringtoString()
             
             Overrides:
             http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
             title="class or interface in java.lang">toStringin 
            classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
            @@ -434,7 +434,7 @@ implements 
             
             getStartKey
            -publicbyte[]getStartKey()
            +publicbyte[]getStartKey()
             
             Specified by:
             getStartKeyin
             interfaceKeyRange
            @@ -447,7 +447,7 @@ implements 
             
             getEndKey
            -publicbyte[]getEndKey()
            +publicbyte[]getEndKey()
             
             Specified by:
             getEndKeyin
             interfaceKeyRange
            @@ -460,7 +460,7 @@ implements 
             
             getTableName
            -publicTableNamegetTableName()
            +publicTableNamegetTableName()
             
             
             
            @@ -469,7 +469,7 @@ implements 
             
             getRegionNameAsString
            -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringgetRegionNameAsString()
            +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringgetRegionNameAsString()
             
             
             
            @@ -478,7 +478,7 @@ implements 
             
             getRegionName
            -publicbyte[]getRegionName()
            +publicbyte[]getRegionName()
             
             
             
            @@ -487,7 +487,7 @@ implements 
             
             getPrimaryHRIForDeployedReplica
            -publicRegionInfogetPrimaryHRIForDeployedReplica()
            +publicRegionInfogetPrimaryHRIForDeployedReplica()
             
             
             
            @@ -496,7 +496,7 @@ implements 
             
             getHdfsRegionDir
            -org.apache.hadoop.fs.PathgetHdfsRegionDir()
            +org.apache.hadoop.fs.PathgetHdfsRegionDir()
             
             
             
            @@ -505,7 +505,7 @@ implements 
             
             containsOnlyHdfsEdits
            -booleancontainsOnlyHdfsEdits()
            +booleancontainsOnlyHdfsEdits()
             
             
             
            @@ -514,7 +514,7 @@ implements 
             
             isHdfsRegioninfoPresent
            -booleanisHdfsRegioninfoPresent()
            +booleanisHdfsRegioninfoPresent()
             
             
             
            @@ -523,7 +523,7 @@ implements 
             
             getModTime
            -longgetModTime()
            +longgetModTime()
             
             
             
            @@ 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/TestTimeout.html
            --
            diff --git a/testdevapidocs/org/apache/hadoop/hbase/TestTimeout.html 
            b/testdevapidocs/org/apache/hadoop/hbase/TestTimeout.html
            index 2847a19..21f4018 100644
            --- a/testdevapidocs/org/apache/hadoop/hbase/TestTimeout.html
            +++ b/testdevapidocs/org/apache/hadoop/hbase/TestTimeout.html
            @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -public class TestTimeout
            +public class TestTimeout
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
             
             
            @@ -130,8 +130,8 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             Field and Description
             
             
            -org.junit.rules.TestRule
            -timeout
            +static HBaseClassTestRule
            +CLASS_RULE
             
             
             
            @@ -197,13 +197,13 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             Field Detail
            -
            +
             
             
             
             
            -timeout
            -public finalorg.junit.rules.TestRule timeout
            +CLASS_RULE
            +public static finalHBaseClassTestRule CLASS_RULE
             
             
             
            @@ -220,7 +220,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             TestTimeout
            -publicTestTimeout()
            +publicTestTimeout()
             
             
             
            @@ -237,7 +237,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             run1
            -publicvoidrun1()
            +publicvoidrun1()
               throws http://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
             title="class or interface in java.lang">InterruptedException
             
             Throws:
            @@ -251,7 +251,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             infiniteLoop
            -publicvoidinfiniteLoop()
            +publicvoidinfiniteLoop()
             Enable to check if timeout works.
              Can't enable as it waits 30seconds and expected doesn't do Exception 
            catching
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/TestZooKeeper.MockLoadBalancer.html
            --
            diff --git 
            a/testdevapidocs/org/apache/hadoop/hbase/TestZooKeeper.MockLoadBalancer.html 
            b/testdevapidocs/org/apache/hadoop/hbase/TestZooKeeper.MockLoadBalancer.html
            index 398dd65..58e2640 100644
            --- a/testdevapidocs/org/apache/hadoop/hbase/TestZooKeeper.MockLoadBalancer.html
            +++ b/testdevapidocs/org/apache/hadoop/hbase/TestZooKeeper.MockLoadBalancer.html
            @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -static class TestZooKeeper.MockLoadBalancer
            +static class TestZooKeeper.MockLoadBalancer
             extends org.apache.hadoop.hbase.master.balancer.SimpleLoadBalancer
             
             
            @@ -266,7 +266,7 @@ extends 
            org.apache.hadoop.hbase.master.balancer.SimpleLoadBalancer
             
             
             retainAssignCalled
            -staticboolean retainAssignCalled
            +staticboolean retainAssignCalled
             
             
             
            @@ -283,7 +283,7 @@ extends 
            org.apache.hadoop.hbase.master.balancer.SimpleLoadBalancer
             
             
             MockLoadBalancer
            -MockLoadBalancer()
            +MockLoadBalancer()
             
             
             
            @@ -300,7 +300,7 @@ extends 
            org.apache.hadoop.hbase.master.balancer.SimpleLoadBalancer
             
             
             retainAssignment
            -publichttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
             title="class or interface in 
            java.util">Maporg.apache.hadoop.hbase.ServerName,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in 
            java.util">Listorg.apache.hadoop.hbase.client.RegionInforetainAssignment(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
             title="class or interface in 
            java.util">Maporg.apache.hadoop.hbase.client.RegionInfo,org.apache.hadoop.hbase.ServerNameregions,
            +publichttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
             title="class or interface in 
            java.util">Maporg.apache.hadoop.hbase.ServerName,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in 
            java.util">Listorg.apache.hadoop.hbase.client.RegionInforetainAssignment(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
             title="class or interface in 
            java.util">Maporg.apache.hadoop.hbase.client.RegionInfo,org.apache.hadoop.hbase.ServerNameregions,
            
             http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in 
            java.util">Listorg.apache.hadoop.hbase.ServerNameservers)
            
              throws org.apache.hadoop.hbase.HBaseIOException
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/TestZooKeeper.html
            --
            diff --git 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
            index eb9e252..667152a 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
            @@ -28,22 +28,22 @@
             020
             021import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
             022import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
            -023import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
            -024import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
            -025import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
            -026import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
            -027import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
            -028import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
            -029import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
            -030import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
            -031import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
            -032import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
            -033import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
            -034import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
            -035import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
            -036import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
            -037import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
            -038import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
            +023import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
            +024import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
            +025import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
            +026import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
            +027import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
            +028import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
            +029import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
            +030import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
            +031import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
            +032import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
            +033import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
            +034import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
            +035import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
            +036import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
            +037import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
            +038import static 
            org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
             039
             040import java.io.IOException;
             041import java.net.URI;
            @@ -70,194 +70,194 @@
             062import 
            org.apache.hadoop.hbase.backup.util.BackupUtils;
             063import 
            org.apache.hadoop.hbase.client.Connection;
             064import 
            org.apache.hadoop.hbase.client.ConnectionFactory;
            -065import 
            org.apache.hbase.thirdparty.com.google.common.collect.Lists;
            -066import 
            org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
            -067import 
            org.apache.yetus.audience.InterfaceAudience;
            -068
            -069/**
            -070 * General backup commands, options and 
            usage messages
            -071 */
            -072
            +065import 
            org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
            +066import 
            org.apache.yetus.audience.InterfaceAudience;
            +067
            +068import 
            org.apache.hbase.thirdparty.com.google.common.collect.Lists;
            +069
            +070/**
            +071 * General backup commands, options and 
            usage messages
            +072 */
             073@InterfaceAudience.Private
             074public final class BackupCommands {
            -075
            -076  public final static String 
            INCORRECT_USAGE = "Incorrect usage";
            -077
            -078  public final static String 
            TOP_LEVEL_NOT_ALLOWED =
            -079  "Top level (root) folder is not 
            allowed to be a backup destination";
            -080
            -081  public static final String USAGE = 
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.Builder.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.Builder.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.Builder.html
            index 914b1c6..03a0b2a 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.Builder.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.Builder.html
            @@ -59,646 +59,711 @@
             051@InterfaceStability.Evolving
             052public class ScannerContext {
             053
            -054  /**
            -055   * Two sets of the same fields. One for 
            the limits, another for the progress towards those limits
            -056   */
            -057  LimitFields limits;
            -058  LimitFields progress;
            -059
            -060  /**
            -061   * The state of the scanner after the 
            invocation of {@link InternalScanner#next(java.util.List)}
            -062   * or {@link 
            RegionScanner#next(java.util.List)}.
            -063   */
            -064  NextState scannerState;
            -065  private static final NextState 
            DEFAULT_STATE = NextState.MORE_VALUES;
            -066
            -067  /**
            -068   * Used as an indication to invocations 
            of {@link InternalScanner#next(java.util.List)} and
            -069   * {@link 
            RegionScanner#next(java.util.List)} that, if true, the progress tracked within 
            this
            -070   * {@link ScannerContext} instance 
            should be considered while evaluating the limits. Useful for
            -071   * enforcing a set of limits across 
            multiple calls (i.e. the limit may not be reached in a single
            -072   * invocation, but any progress made 
            should be considered in future invocations)
            -073   * p
            -074   * Defaulting this value to false means 
            that, by default, any tracked progress will be wiped clean
            -075   * on invocations to {@link 
            InternalScanner#next(java.util.List)} and
            -076   * {@link 
            RegionScanner#next(java.util.List)} and the call will be treated as though no 
            progress
            -077   * has been made towards the limits so 
            far.
            -078   * p
            -079   * This is an important mechanism. 
            Users of Internal/Region scanners expect that they can define
            -080   * some limits and then repeatedly 
            invoke {@link InternalScanner#next(List)} or
            -081   * {@link RegionScanner#next(List)} 
            where each invocation respects these limits separately.
            -082   * p
            -083   * For example: pre {@code
            -084   * ScannerContext context = new 
            ScannerContext.newBuilder().setBatchLimit(5).build();
            -085   * RegionScanner scanner = ...
            -086   * ListCell results = new 
            ArrayListCell();
            -087   * while(scanner.next(results, 
            context)) {
            -088   *   // Do something with a batch of 5 
            cells
            -089   * }
            -090   * }/pre However, in the case 
            of RPCs, the server wants to be able to define a set of
            -091   * limits for a particular RPC request 
            and have those limits respected across multiple
            -092   * invocations. This means that the 
            progress made towards the limits in earlier calls will be
            -093   * saved and considered in future 
            invocations
            -094   */
            -095  boolean keepProgress;
            -096  private static boolean 
            DEFAULT_KEEP_PROGRESS = false;
            -097
            -098  private Cell lastPeekedCell = null;
            +054  LimitFields limits;
            +055  /**
            +056   * A different set of progress fields. 
            Only include batch, dataSize and heapSize. Compare to
            +057   * LimitFields, ProgressFields doesn't 
            contain time field. As we save a deadline in LimitFields,
            +058   * so use {@link 
            System#currentTimeMillis()} directly when check time limit.
            +059   */
            +060  ProgressFields progress;
            +061
            +062  /**
            +063   * The state of the scanner after the 
            invocation of {@link InternalScanner#next(java.util.List)}
            +064   * or {@link 
            RegionScanner#next(java.util.List)}.
            +065   */
            +066  NextState scannerState;
            +067  private static final NextState 
            DEFAULT_STATE = NextState.MORE_VALUES;
            +068
            +069  /**
            +070   * Used as an indication to invocations 
            of {@link InternalScanner#next(java.util.List)} and
            +071   * {@link 
            RegionScanner#next(java.util.List)} that, if true, the progress tracked within 
            this
            +072   * {@link ScannerContext} instance 
            should be considered while evaluating the limits. Useful for
            +073   * enforcing a set of limits across 
            multiple calls (i.e. the limit may not be reached in a single
            +074   * invocation, but any progress made 
            should be considered in future invocations)
            +075   * p
            +076   * Defaulting this value to false means 
            that, by default, any tracked progress will be wiped clean
            +077   * on invocations to {@link 
            InternalScanner#next(java.util.List)} and
            +078   * {@link 
            RegionScanner#next(java.util.List)} and the call will be treated as though no 
            progress
            +079   * has been made towards the limits so 
            far.
            +080   * p
            +081   * This is an important mechanism. 
            Users of Internal/Region scanners expect that they can define
            +082   * some limits and then repeatedly 
            invoke {@link InternalScanner#next(List)} or
            +083   * {@link RegionScanner#next(List)} 
            where each invocation respects these limits separately.
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntryGroup.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntryGroup.html
             
            b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntryGroup.html
            index 9baa74f..c6877be 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntryGroup.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntryGroup.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10};
            +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10};
             var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private class BucketCache.BucketEntryGroup
            +private class BucketCache.BucketEntryGroup
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
             implements http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
             title="class or interface in java.lang">ComparableBucketCache.BucketEntryGroup
             Used to group bucket entries into priority buckets. There 
            will be a
            @@ -198,17 +198,21 @@ implements http://docs.oracle.com/javase/8/docs/api/java/lang/Comparabl
             
             
             boolean
            -equals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Objectthat)
            +equals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Objecto)
             
             
             long
             free(longtoFree)
             
             
            +int
            +hashCode()
            +
            +
             long
             overflow()
             
            -
            +
             long
             totalSize()
             
            @@ -218,7 +222,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/lang/Comparabl
             
             
             Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
            -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
             title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
             title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
             title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
             title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
             title="class or interface in java.lang">notify, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notifyAll--;
             title="class or interface in java.lang">notifyAll, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.ht
             ml?is-external=true#toString--" title="class or interface in 
            java.lang">toString, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--;
             title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-;
             title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-;
             title="class or interface in java.lang">wait
            +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
             title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
             title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
             title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
             title="class or interface in java.lang">notify, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notifyAll--;
             title="class or interface in java.lang">notifyAll, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
             title="class or interface in java.lang">toString, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.ht
             ml?is-external=true#wait--" title="class or interface in java.lang">wait, 
            http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-;
             title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-;
             title="class or interface in java.lang">wait
             
             
             
            @@ -240,7 +244,7 @@ implements 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html
            index e05510e..2e114d8 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html
            @@ -54,891 +54,884 @@
             046import 
            org.apache.hadoop.hbase.HTableDescriptor;
             047import 
            org.apache.hadoop.hbase.MetaTableAccessor;
             048import 
            org.apache.hadoop.hbase.MetaTableAccessor.DefaultVisitorBase;
            -049import 
            org.apache.hadoop.hbase.NamespaceDescriptor;
            -050import 
            org.apache.hadoop.hbase.ServerName;
            -051import 
            org.apache.hadoop.hbase.TableName;
            -052import 
            org.apache.hadoop.hbase.client.ClusterConnection;
            -053import 
            org.apache.hadoop.hbase.client.Delete;
            -054import 
            org.apache.hadoop.hbase.client.Get;
            -055import 
            org.apache.hadoop.hbase.client.Mutation;
            -056import 
            org.apache.hadoop.hbase.client.Put;
            -057import 
            org.apache.hadoop.hbase.client.RegionInfo;
            -058import 
            org.apache.hadoop.hbase.client.Result;
            -059import 
            org.apache.hadoop.hbase.client.Scan;
            -060import 
            org.apache.hadoop.hbase.client.Table;
            -061import 
            org.apache.hadoop.hbase.client.TableState;
            -062import 
            org.apache.hadoop.hbase.constraint.ConstraintException;
            -063import 
            org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint;
            -064import 
            org.apache.hadoop.hbase.exceptions.DeserializationException;
            -065import 
            org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
            -066import 
            org.apache.hadoop.hbase.master.MasterServices;
            -067import 
            org.apache.hadoop.hbase.master.ServerListener;
            -068import 
            org.apache.hadoop.hbase.master.TableStateManager;
            -069import 
            org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode;
            -070import 
            org.apache.hadoop.hbase.net.Address;
            -071import 
            org.apache.hadoop.hbase.procedure2.Procedure;
            -072import 
            org.apache.hadoop.hbase.protobuf.ProtobufMagic;
            -073import 
            org.apache.hadoop.hbase.protobuf.ProtobufUtil;
            -074import 
            org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos;
            -075import 
            org.apache.hadoop.hbase.protobuf.generated.RSGroupProtos;
            -076import 
            org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy;
            -077import 
            org.apache.hadoop.hbase.security.access.AccessControlLists;
            -078import 
            org.apache.hadoop.hbase.util.Bytes;
            -079import 
            org.apache.hadoop.hbase.zookeeper.ZKUtil;
            -080import 
            org.apache.hadoop.hbase.zookeeper.ZKWatcher;
            -081import 
            org.apache.hadoop.hbase.zookeeper.ZNodePaths;
            -082import 
            org.apache.yetus.audience.InterfaceAudience;
            -083import 
            org.apache.zookeeper.KeeperException;
            -084import org.slf4j.Logger;
            -085import org.slf4j.LoggerFactory;
            -086
            -087import 
            org.apache.hbase.thirdparty.com.google.common.collect.Lists;
            -088import 
            org.apache.hbase.thirdparty.com.google.common.collect.Maps;
            -089import 
            org.apache.hbase.thirdparty.com.google.common.collect.Sets;
            -090import 
            org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
            -091import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
            -092
            -093/**
            -094 * This is an implementation of {@link 
            RSGroupInfoManager} which makes
            -095 * use of an HBase table as the 
            persistence store for the group information.
            -096 * It also makes use of zookeeper to 
            store group information needed
            -097 * for bootstrapping during offline 
            mode.
            -098 *
            -099 * h2Concurrency/h2
            -100 * RSGroup state is kept locally in Maps. 
            There is a rsgroup name to cached
            -101 * RSGroupInfo Map at {@link #rsGroupMap} 
            and a Map of tables to the name of the
            -102 * rsgroup they belong too (in {@link 
            #tableMap}). These Maps are persisted to the
            -103 * hbase:rsgroup table (and cached in zk) 
            on each modification.
            -104 *
            -105 * pMutations on state are 
            synchronized but reads can continue without having
            -106 * to wait on an instance monitor, 
            mutations do wholesale replace of the Maps on
            -107 * update -- Copy-On-Write; the local 
            Maps of state are read-only, just-in-case
            -108 * (see flushConfig).
            -109 *
            -110 * pReads must not block else 
            there is a danger we'll deadlock.
            -111 *
            -112 * pClients of this class, the 
            {@link RSGroupAdminEndpoint} for example, want to query and
            -113 * then act on the results of the query 
            modifying cache in zookeeper without another thread
            -114 * making intermediate modifications. 
            These clients synchronize on the 'this' instance so
            -115 * no other has access concurrently. 
            Reads must be able to continue concurrently.
            -116 */
            -117@InterfaceAudience.Private
            -118final class RSGroupInfoManagerImpl 
            implements RSGroupInfoManager {
            -119  private static final Logger LOG = 
            LoggerFactory.getLogger(RSGroupInfoManagerImpl.class);
            -120
            -121  /** Table descriptor for 
            codehbase:rsgroup/code catalog table */
            -122  private final 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/integration.html
            --
            diff --git a/hbase-build-configuration/integration.html 
            b/hbase-build-configuration/integration.html
            index e2dd2e2..1a72d4d 100644
            --- a/hbase-build-configuration/integration.html
            +++ b/hbase-build-configuration/integration.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Build Configuration  CI Management
             
            @@ -126,7 +126,7 @@
             https://www.apache.org/;>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/issue-tracking.html
            --
            diff --git a/hbase-build-configuration/issue-tracking.html 
            b/hbase-build-configuration/issue-tracking.html
            index 0062a19..c108358 100644
            --- a/hbase-build-configuration/issue-tracking.html
            +++ b/hbase-build-configuration/issue-tracking.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Build Configuration  Issue Management
             
            @@ -123,7 +123,7 @@
             https://www.apache.org/;>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/license.html
            --
            diff --git a/hbase-build-configuration/license.html 
            b/hbase-build-configuration/license.html
            index 989d0f7..6ec28ca 100644
            --- a/hbase-build-configuration/license.html
            +++ b/hbase-build-configuration/license.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Build Configuration  Project Licenses
             
            @@ -326,7 +326,7 @@
             https://www.apache.org/;>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/mail-lists.html
            --
            diff --git a/hbase-build-configuration/mail-lists.html 
            b/hbase-build-configuration/mail-lists.html
            index 8e37d21..4722133 100644
            --- a/hbase-build-configuration/mail-lists.html
            +++ b/hbase-build-configuration/mail-lists.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Build Configuration  Project Mailing 
            Lists
             
            @@ -176,7 +176,7 @@
             https://www.apache.org/;>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/plugin-management.html
            --
            diff --git a/hbase-build-configuration/plugin-management.html 
            b/hbase-build-configuration/plugin-management.html
            index 34048dc..6ef 100644
            --- a/hbase-build-configuration/plugin-management.html
            +++ b/hbase-build-configuration/plugin-management.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Build Configuration  Project Plugin 
            Management
             
            @@ -271,7 +271,7 @@
             https://www.apache.org/;>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/plugins.html
            --
            diff --git a/hbase-build-configuration/plugins.html 
            b/hbase-build-configuration/plugins.html
            index e4617dc..3b0cda5 100644
            --- a/hbase-build-configuration/plugins.html
            +++ b/hbase-build-configuration/plugins.html
            @@ -7,7 +7,7 @@
               
             
             
            -
            +
             
             Apache HBase - Build Configuration  Project Plugins
             
            @@ -214,7 +214,7 @@
             https://www.apache.org/;>The Apache Software 
            Foundation.
             All rights reserved.  
             
            -  Last Published: 
            2018-01-11
            +  Last Published: 
            2018-01-12
             
             
             
            
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.RegionGroupingStrategy.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.RegionGroupingStrategy.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.RegionGroupingStrategy.html
            index 4fec5eb..63c9ca7 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.RegionGroupingStrategy.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.RegionGroupingStrategy.html
            @@ -35,254 +35,262 @@
             027import java.util.List;
             028import 
            java.util.concurrent.ConcurrentHashMap;
             029import 
            java.util.concurrent.ConcurrentMap;
            -030
            +030import java.util.concurrent.locks.Lock;
             031import 
            org.apache.hadoop.conf.Configuration;
            -032import 
            org.apache.yetus.audience.InterfaceAudience;
            -033import org.slf4j.Logger;
            -034import org.slf4j.LoggerFactory;
            -035// imports for classes still in 
            regionserver.wal
            -036import 
            org.apache.hadoop.hbase.regionserver.wal.WALActionsListener;
            -037import 
            org.apache.hadoop.hbase.util.Bytes;
            -038import 
            org.apache.hadoop.hbase.util.IdLock;
            -039
            -040/**
            -041 * A WAL Provider that returns a WAL per 
            group of regions.
            -042 *
            -043 * This provider follows the decorator 
            pattern and mainly holds the logic for WAL grouping.
            -044 * WAL creation/roll/close is delegated 
            to {@link #DELEGATE_PROVIDER}
            -045 *
            -046 * Region grouping is handled via {@link 
            RegionGroupingStrategy} and can be configured via the
            -047 * property 
            "hbase.wal.regiongrouping.strategy". Current strategy choices are
            -048 * ul
            -049 *   
            liemdefaultStrategy/em : Whatever strategy this version 
            of HBase picks. currently
            -050 *  
            "bounded"./li
            -051 *   
            liemidentity/em : each region belongs to its own 
            group./li
            -052 *   
            liembounded/em : bounded number of groups and region 
            evenly assigned to each group./li
            -053 * /ul
            -054 * Optionally, a FQCN to a custom 
            implementation may be given.
            -055 */
            -056@InterfaceAudience.Private
            -057public class RegionGroupingProvider 
            implements WALProvider {
            -058  private static final Logger LOG = 
            LoggerFactory.getLogger(RegionGroupingProvider.class);
            -059
            -060  /**
            -061   * Map identifiers to a group number.
            -062   */
            -063  public static interface 
            RegionGroupingStrategy {
            -064String GROUP_NAME_DELIMITER = ".";
            -065
            -066/**
            -067 * Given an identifier and a 
            namespace, pick a group.
            -068 */
            -069String group(final byte[] identifier, 
            byte[] namespace);
            -070void init(Configuration config, 
            String providerId);
            -071  }
            -072
            -073  /**
            -074   * Maps between configuration names for 
            strategies and implementation classes.
            -075   */
            -076  static enum Strategies {
            -077
            defaultStrategy(BoundedGroupingStrategy.class),
            -078
            identity(IdentityGroupingStrategy.class),
            -079
            bounded(BoundedGroupingStrategy.class),
            -080
            namespace(NamespaceGroupingStrategy.class);
            -081
            -082final Class? extends 
            RegionGroupingStrategy clazz;
            -083Strategies(Class? extends 
            RegionGroupingStrategy clazz) {
            -084  this.clazz = clazz;
            -085}
            -086  }
            -087
            -088  /**
            -089   * instantiate a strategy from a config 
            property.
            -090   * requires conf to have already been 
            set (as well as anything the provider might need to read).
            -091   */
            -092  RegionGroupingStrategy 
            getStrategy(final Configuration conf, final String key,
            -093  final String defaultValue) throws 
            IOException {
            -094Class? extends 
            RegionGroupingStrategy clazz;
            -095try {
            -096  clazz = 
            Strategies.valueOf(conf.get(key, defaultValue)).clazz;
            -097} catch (IllegalArgumentException 
            exception) {
            -098  // Fall back to them specifying a 
            class name
            -099  // Note that the passed default 
            class shouldn't actually be used, since the above only fails
            -100  // when there is a config value 
            present.
            -101  clazz = conf.getClass(key, 
            IdentityGroupingStrategy.class, RegionGroupingStrategy.class);
            -102}
            -103LOG.info("Instantiating 
            RegionGroupingStrategy of type " + clazz);
            -104try {
            -105  final RegionGroupingStrategy result 
            = clazz.newInstance();
            -106  result.init(conf, providerId);
            -107  return result;
            -108} catch (InstantiationException 
            exception) {
            -109  LOG.error("couldn't set up region 
            grouping strategy, check config key " +
            -110  REGION_GROUPING_STRATEGY);
            -111  LOG.debug("Exception details for 
            failure to load region grouping strategy.", exception);
            -112  throw new IOException("couldn't set 
            up region grouping strategy", exception);
            -113} catch (IllegalAccessException 
            exception) {
            -114  LOG.error("couldn't set up region 
            grouping strategy, check config key " +
            -115  REGION_GROUPING_STRATEGY);
            -116  LOG.debug("Exception details for 
            failure to load region grouping strategy.", exception);
            -117   

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
            new file mode 100644
            index 000..960f2b4
            --- /dev/null
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
            @@ -0,0 +1,368 @@
            +http://www.w3.org/TR/html4/loose.dtd;>
            +
            +
            +
            +
            +
            +HBaseAdmin.ReplicationFuture (Apache HBase 3.0.0-SNAPSHOT API)
            +
            +
            +
            +
            +
            +var methods = {"i0":10};
            +var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
            +var altColor = "altColor";
            +var rowColor = "rowColor";
            +var tableTab = "tableTab";
            +var activeTableTab = "activeTableTab";
            +
            +
            +JavaScript is disabled on your browser.
            +
            +
            +
            +
            +
            +Skip navigation links
            +
            +
            +
            +
            +Overview
            +Package
            +Class
            +Use
            +Tree
            +Deprecated
            +Index
            +Help
            +
            +
            +
            +
            +PrevClass
            +NextClass
            +
            +
            +Frames
            +NoFrames
            +
            +
            +AllClasses
            +
            +
            +
            +
            +
            +
            +
            +Summary:
            +Nested|
            +Field|
            +Constr|
            +Method
            +
            +
            +Detail:
            +Field|
            +Constr|
            +Method
            +
            +
            +
            +
            +
            +
            +
            +
            +org.apache.hadoop.hbase.client
            +Class 
            HBaseAdmin.ReplicationFuture
            +
            +
            +
            +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">java.lang.Object
            +
            +
            +org.apache.hadoop.hbase.client.HBaseAdmin.ProcedureFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            +
            +
            +org.apache.hadoop.hbase.client.HBaseAdmin.ReplicationFuture
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +All Implemented Interfaces:
            +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Future.html?is-external=true;
             title="class or interface in java.util.concurrent">Futurehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            +
            +
            +Enclosing class:
            +HBaseAdmin
            +
            +
            +
            +@InterfaceAudience.Private
            + @InterfaceStability.Evolving
            +private static class HBaseAdmin.ReplicationFuture
            +extends HBaseAdmin.ProcedureFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
             title="class or interface in java.lang">Void
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +Nested Class Summary
            +
            +
            +
            +
            +Nested classes/interfaces inherited from 
            classorg.apache.hadoop.hbase.client.HBaseAdmin.ProcedureFuture
            +HBaseAdmin.ProcedureFuture.WaitForStateCallable
            +
            +
            +
            +
            +
            +
            +
            +
            +Field Summary
            +
            +Fields
            +
            +Modifier and Type
            +Field and Description
            +
            +
            +private http://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true;
             title="class or interface in java.util.function">Supplierhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String
            +getOperation
            +
            +
            +private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String
            +peerId
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +Constructor Summary
            +
            +Constructors
            +
            +Constructor and Description
            +
            +
            +ReplicationFuture(HBaseAdminadmin,
            + http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringpeerId,
            + http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
             title="class or interface in java.lang">LongprocId,
            + http://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true;
             title="class or interface in java.util.function">Supplierhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in 
            java.lang">StringgetOperation)
            +
            +
            +
            +
            +
            +
            +
            +
            +
            +Method Summary
            +
            +All MethodsInstance MethodsConcrete Methods
            +
            +Modifier and Type
            +Method and Description
            +
            +
            +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String
            +toString()
            +
            +
            +
            +
            +
            +
            +Methods inherited from classorg.apache.hadoop.hbase.client.HBaseAdmin.ProcedureFuture
            +abortProcedureResult,
             cancel,
             convertResult,
             get,
             get,
             getAdmin,
             getProcedureResult,
             isCancelled,
             isDone,
             postOperationFailure,
             postOperationResult,
             waitForState,
             waitOperationResu
             lt
            +
            +
            +
            +
            +
            +Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html 
            b/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
            index dd4afb1..c8dd758 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = {"i0":10,"i1":10,"i2":10,"i3":10};
            +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10};
             var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
             
             
             @InterfaceAudience.Private
            -public class TableSnapshotScanner
            +public class TableSnapshotScanner
             extends AbstractClientScanner
             A Scanner which performs a scan over snapshot files. Using 
            this class requires copying the
              snapshot to a temporary empty directory, which will copy the snapshot 
            reference files into that
            @@ -205,6 +205,10 @@ extends scan
             
             
            +private boolean
            +snapshotAlreadyRestored
            +
            +
             private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String
             snapshotName
             
            @@ -234,11 +238,19 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringsnapshotName,
            -Scanscan)
            +Scanscan)
            +
            +
            +TableSnapshotScanner(org.apache.hadoop.conf.Configurationconf,
            +org.apache.hadoop.fs.PathrootDir,
            +org.apache.hadoop.fs.PathrestoreDir,
            +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringsnapshotName,
            +Scanscan,
            +booleansnapshotAlreadyRestored)
             Creates a TableSnapshotScanner.
             
             
            -
            +
             TableSnapshotScanner(org.apache.hadoop.conf.Configurationconf,
             org.apache.hadoop.fs.PathrestoreDir,
             http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringsnapshotName,
            @@ -262,22 +274,34 @@ extends Method and Description
             
             
            +private void
            +cleanup()
            +
            +
             void
             close()
             Closes the scanner and releases any resources it has 
            allocated
             
             
            -
            -private void
            -init()
            -
             
            +private boolean
            +isValidRegion(RegionInfohri)
            +
            +
             Result
             next()
             Grab the next row's worth of values.
             
             
            -
            +
            +private void
            +openWithoutRestoringSnapshot()
            +
            +
            +private void
            +openWithRestoringSnapshot()
            +
            +
             boolean
             renewLease()
             Allow the client to renew the scanner's lease on the 
            server.
            @@ -332,7 +356,7 @@ extends 
             
             LOG
            -private static finalorg.slf4j.Logger LOG
            +private static finalorg.slf4j.Logger LOG
             
             
             
            @@ -341,7 +365,7 @@ extends 
             
             conf
            -privateorg.apache.hadoop.conf.Configuration conf
            +privateorg.apache.hadoop.conf.Configuration conf
             
             
             
            @@ -350,7 +374,7 @@ extends 
             
             snapshotName
            -privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String snapshotName
            +privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String snapshotName
             
             
             
            @@ -359,7 +383,7 @@ extends 
             
             fs
            -privateorg.apache.hadoop.fs.FileSystem fs
            +privateorg.apache.hadoop.fs.FileSystem fs
             
             
             
            @@ -368,7 +392,7 @@ extends 
             
             rootDir
            -privateorg.apache.hadoop.fs.Path rootDir
            +privateorg.apache.hadoop.fs.Path rootDir
             
             
             
            @@ -377,7 +401,7 @@ extends 
             
             restoreDir
            -privateorg.apache.hadoop.fs.Path restoreDir
            +privateorg.apache.hadoop.fs.Path restoreDir
             
             
             
            @@ -386,7 +410,7 @@ extends 
             
             scan
            -privateScan scan
            +privateScan scan
             
             
             
            @@ -395,7 +419,7 @@ extends 
             
             regions
            -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true;
             title="class or interface in java.util">ArrayListRegionInfo regions
            +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true;
             title="class or interface in java.util">ArrayListRegionInfo regions
             
             
             
            @@ -404,7 +428,16 @@ extends 
             
             htd
            -privateTableDescriptor htd
            +privateTableDescriptor htd
            +
            +
            +
            +
            +
            +
            +
            +snapshotAlreadyRestored
            +private finalboolean snapshotAlreadyRestored
             
             
             
            @@ -413,7 +446,7 @@ extends 
             
             currentRegionScanner
            -privateClientSideRegionScanner currentRegionScanner
            +privateClientSideRegionScanner currentRegionScanner
             
             
             
            @@ -422,7 +455,7 @@ extends 
             
             currentRegion
            -privateint currentRegion
            +privateint currentRegion
             
             
             
            @@ -439,7 +472,7 @@ extends 
             
             TableSnapshotScanner
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html
            index 5b3b750..a1f3f7e 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html
            @@ -97,3307 +97,3304 @@
             089import 
            org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer;
             090import 
            org.apache.hbase.thirdparty.io.netty.util.Timeout;
             091import 
            org.apache.hbase.thirdparty.io.netty.util.TimerTask;
            -092import 
            org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
            -093import 
            org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
            -094import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
            -095import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
            -096import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse;
            -097import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
            -098import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse;
            -099import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
            -100import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse;
            -101import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest;
            -102import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse;
            -103import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
            -104import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
            -105import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest;
            -106import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse;
            -107import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest;
            -108import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse;
            -109import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest;
            -110import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse;
            -111import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest;
            -112import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse;
            -113import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
            -114import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
            -115import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
            -116import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
            -117import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
            -118import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
            -119import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
            -120import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
            -121import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
            -122import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
            -123import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
            -124import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest;
            -125import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse;
            -126import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
            -127import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
            -128import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest;
            -129import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse;
            -130import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest;
            -131import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse;
            -132import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
            -133import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
            -134import 
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState.html
            index 0ef5497..79102c5 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState.html
            @@ -188,6 +188,6 @@ the order they are declared.
             
             
             
            -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
            +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerImpl.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerImpl.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerImpl.html
            index 0d1a0b4..1d27530 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerImpl.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerImpl.html
            @@ -162,6 +162,6 @@
             
             
             
            -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
            +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState.html
            index f8abbbe..c30fde4 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState.html
            @@ -184,6 +184,6 @@ the order they are declared.
             
             
             
            -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
            +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.html
            index ae89acd..38aaa09 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncScanSingleRegionRpcRetryingCaller.html
            @@ -162,6 +162,6 @@
             
             
             
            -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
            +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncServerRequestRpcRetryingCaller.Callable.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncServerRequestRpcRetryingCaller.Callable.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncServerRequestRpcRetryingCaller.Callable.html
            index 34d6483..772b83d 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncServerRequestRpcRetryingCaller.Callable.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncServerRequestRpcRetryingCaller.Callable.html
            @@ -198,6 +198,6 @@
             
             
             
            -Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
            +Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
            reserved.
             
             
            
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionWorker.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionWorker.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionWorker.html
            index 91eec45..d1cd185 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionWorker.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionWorker.html
            @@ -88,400 +88,396 @@
             080  private final static String 
            CONF_COMPACT_ONCE = "hbase.compactiontool.compact.once";
             081  private final static String 
            CONF_COMPACT_MAJOR = "hbase.compactiontool.compact.major";
             082  private final static String 
            CONF_DELETE_COMPACTED = "hbase.compactiontool.delete";
            -083  private final static String 
            CONF_COMPLETE_COMPACTION = "hbase.hstore.compaction.complete";
            -084
            -085  /**
            -086   * Class responsible to execute the 
            Compaction on the specified path.
            -087   * The path can be a table, region or 
            family directory.
            -088   */
            -089  private static class CompactionWorker 
            {
            -090private final boolean 
            keepCompactedFiles;
            -091private final boolean 
            deleteCompacted;
            -092private final Configuration conf;
            -093private final FileSystem fs;
            -094private final Path tmpDir;
            -095
            -096public CompactionWorker(final 
            FileSystem fs, final Configuration conf) {
            -097  this.conf = conf;
            -098  this.keepCompactedFiles = 
            !conf.getBoolean(CONF_COMPLETE_COMPACTION, true);
            -099  this.deleteCompacted = 
            conf.getBoolean(CONF_DELETE_COMPACTED, false);
            -100  this.tmpDir = new 
            Path(conf.get(CONF_TMP_DIR));
            -101  this.fs = fs;
            -102}
            -103
            -104/**
            -105 * Execute the compaction on the 
            specified path.
            -106 *
            -107 * @param path Directory path on 
            which to run compaction.
            -108 * @param compactOnce Execute just a 
            single step of compaction.
            -109 * @param major Request major 
            compaction.
            -110 */
            -111public void compact(final Path path, 
            final boolean compactOnce, final boolean major) throws IOException {
            -112  if (isFamilyDir(fs, path)) {
            -113Path regionDir = 
            path.getParent();
            -114Path tableDir = 
            regionDir.getParent();
            -115TableDescriptor htd = 
            FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir);
            -116RegionInfo hri = 
            HRegionFileSystem.loadRegionInfoFileContent(fs, regionDir);
            -117compactStoreFiles(tableDir, htd, 
            hri,
            -118path.getName(), compactOnce, 
            major);
            -119  } else if (isRegionDir(fs, path)) 
            {
            -120Path tableDir = 
            path.getParent();
            -121TableDescriptor htd = 
            FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir);
            -122compactRegion(tableDir, htd, 
            path, compactOnce, major);
            -123  } else if (isTableDir(fs, path)) 
            {
            -124compactTable(path, compactOnce, 
            major);
            -125  } else {
            -126throw new IOException(
            -127  "Specified path is not a table, 
            region or family directory. path=" + path);
            -128  }
            -129}
            -130
            -131private void compactTable(final Path 
            tableDir, final boolean compactOnce, final boolean major)
            -132throws IOException {
            -133  TableDescriptor htd = 
            FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir);
            -134  for (Path regionDir: 
            FSUtils.getRegionDirs(fs, tableDir)) {
            -135compactRegion(tableDir, htd, 
            regionDir, compactOnce, major);
            -136  }
            -137}
            -138
            -139private void compactRegion(final Path 
            tableDir, final TableDescriptor htd,
            -140final Path regionDir, final 
            boolean compactOnce, final boolean major)
            -141throws IOException {
            -142  RegionInfo hri = 
            HRegionFileSystem.loadRegionInfoFileContent(fs, regionDir);
            -143  for (Path familyDir: 
            FSUtils.getFamilyDirs(fs, regionDir)) {
            -144compactStoreFiles(tableDir, htd, 
            hri, familyDir.getName(), compactOnce, major);
            -145  }
            -146}
            -147
            -148/**
            -149 * Execute the actual compaction 
            job.
            -150 * If the compact once flag is not 
            specified, execute the compaction until
            -151 * no more compactions are needed. 
            Uses the Configuration settings provided.
            -152 */
            -153private void compactStoreFiles(final 
            Path tableDir, final TableDescriptor htd,
            -154final RegionInfo hri, final 
            String familyName, final boolean compactOnce,
            -155final boolean major) throws 
            IOException {
            -156  HStore store = getStore(conf, fs, 
            tableDir, htd, hri, familyName, tmpDir);
            -157  LOG.info("Compact table=" + 
            htd.getTableName() +
            -158" region=" + 
            hri.getRegionNameAsString() +
            -159" family=" + familyName);
            -160  if (major) {
            -161store.triggerMajorCompaction();
            -162  }
            -163  do {
            -164OptionalCompactionContext 
            compaction =
            -165
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html 
            b/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
            index 650d5d3..f65cdb2 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
            @@ -274,7 +274,7 @@ the order they are declared.
             
             
             values
            -public staticTableState.State[]values()
            +public staticTableState.State[]values()
             Returns an array containing the constants of this enum 
            type, in
             the order they are declared.  This method may be used to iterate
             over the constants as follows:
            @@ -294,7 +294,7 @@ for (TableState.State c : TableState.State.values())
             
             
             valueOf
            -public staticTableState.StatevalueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringname)
            +public staticTableState.StatevalueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringname)
             Returns the enum constant of this type with the specified 
            name.
             The string must match exactly an identifier used to declare an
             enum constant in this type.  (Extraneous whitespace characters are 
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/client/class-use/Admin.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Admin.html 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/Admin.html
            index c0cf2ff..430067c 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Admin.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Admin.html
            @@ -422,7 +422,7 @@ service.
             
             
             
            -(package private) 
            org.apache.hadoop.hbase.shaded.com.google.common.collect.MultimapTableName,http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String
            +(package private) 
            org.apache.hbase.thirdparty.com.google.common.collect.MultimapTableName,http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String
             SnapshotQuotaObserverChore.getSnapshotsFromTables(Adminadmin,
               http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
             title="class or interface in java.util">SetTableNametablesToFetchSnapshotsFrom)
             Computes a mapping of originating TableName to 
            snapshots, when the TableName
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/client/class-use/AdvancedScanResultConsumer.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AdvancedScanResultConsumer.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AdvancedScanResultConsumer.html
            index 28a8c94..5cb2039 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AdvancedScanResultConsumer.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AdvancedScanResultConsumer.html
            @@ -292,7 +292,7 @@
               intstartLogErrorsCnt)
             
             
            -AsyncScanSingleRegionRpcRetryingCaller(org.apache.hadoop.hbase.shaded.io.netty.util.HashedWheelTimerretryTimer,
            +AsyncScanSingleRegionRpcRetryingCaller(org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimerretryTimer,
               AsyncConnectionImplconn,
               Scanscan,
               ScanMetricsscanMetrics,
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdminBuilderBase.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdminBuilderBase.html
             
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdminBuilderBase.html
            index abf11c4..8638271 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdminBuilderBase.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdminBuilderBase.html
            @@ -104,8 +104,8 @@
             
             
             
            -RawAsyncHBaseAdmin(AsyncConnectionImplconnection,
            -  
            org.apache.hadoop.hbase.shaded.io.netty.util.HashedWheelTimerretryTimer,
            +RawAsyncHBaseAdmin(AsyncConnectionImplconnection,
            +  
            org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimerretryTimer,
               AsyncAdminBuilderBasebuilder)
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdminRequestRetryingCaller.Callable.html
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.ValueAndTagRewriteByteBufferCell.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.ValueAndTagRewriteByteBufferCell.html
             
            b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.ValueAndTagRewriteByteBufferCell.html
            deleted file mode 100644
            index 1dc3d76..000
            --- 
            a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.ValueAndTagRewriteByteBufferCell.html
            +++ /dev/null
            @@ -1,621 +0,0 @@
            -http://www.w3.org/TR/html4/loose.dtd;>
            -
            -
            -
            -
            -
            -PrivateCellUtil.ValueAndTagRewriteByteBufferCell (Apache HBase 
            3.0.0-SNAPSHOT API)
            -
            -
            -
            -
            -
            -var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10};
            -var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
            -var altColor = "altColor";
            -var rowColor = "rowColor";
            -var tableTab = "tableTab";
            -var activeTableTab = "activeTableTab";
            -
            -
            -JavaScript is disabled on your browser.
            -
            -
            -
            -
            -
            -Skip navigation links
            -
            -
            -
            -
            -Overview
            -Package
            -Class
            -Use
            -Tree
            -Deprecated
            -Index
            -Help
            -
            -
            -
            -
            -PrevClass
            -NextClass
            -
            -
            -Frames
            -NoFrames
            -
            -
            -AllClasses
            -
            -
            -
            -
            -
            -
            -
            -Summary:
            -Nested|
            -Field|
            -Constr|
            -Method
            -
            -
            -Detail:
            -Field|
            -Constr|
            -Method
            -
            -
            -
            -
            -
            -
            -
            -
            -org.apache.hadoop.hbase
            -Class PrivateCellUtil.ValueAndTagRewriteByteBufferCell
            -
            -
            -
            -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">java.lang.Object
            -
            -
            -org.apache.hadoop.hbase.ByteBufferCell
            -
            -
            -org.apache.hadoop.hbase.PrivateCellUtil.TagRewriteByteBufferCell
            -
            -
            -org.apache.hadoop.hbase.PrivateCellUtil.ValueAndTagRewriteByteBufferCell
            -
            -
            -
            -
            -
            -
            -
            -
            -
            -
            -
            -All Implemented Interfaces:
            -http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true;
             title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell
            -
            -
            -Enclosing class:
            -PrivateCellUtil
            -
            -
            -
            -static class PrivateCellUtil.ValueAndTagRewriteByteBufferCell
            -extends PrivateCellUtil.TagRewriteByteBufferCell
            -
            -
            -
            -
            -
            -
            -
            -
            -
            -
            -
            -Nested Class Summary
            -
            -
            -
            -
            -Nested classes/interfaces inherited from 
            interfaceorg.apache.hadoop.hbase.Cell
            -Cell.DataType
            -
            -
            -
            -
            -
            -
            -
            -
            -Field Summary
            -
            -Fields
            -
            -Modifier and Type
            -Field and Description
            -
            -
            -protected byte[]
            -value
            -
            -
            -
            -
            -
            -
            -Fields inherited from classorg.apache.hadoop.hbase.PrivateCellUtil.TagRewriteByteBufferCell
            -cell,
             tags
            -
            -
            -
            -
            -
            -Fields inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell
            -CELL_NOT_BASED_ON_CHUNK
            -
            -
            -
            -
            -
            -Fields inherited from interfaceorg.apache.hadoop.hbase.RawCell
            -MAX_TAGS_LENGTH
            -
            -
            -
            -
            -
            -
            -
            -
            -Constructor Summary
            -
            -Constructors
            -
            -Constructor and Description
            -
            -
            -ValueAndTagRewriteByteBufferCell(ByteBufferCellcell,
            -byte[]value,
            -byte[]tags)
            -
            -
            -
            -
            -
            -
            -
            -
            -
            -Method Summary
            -
            -All MethodsInstance MethodsConcrete Methods
            -
            -Modifier and Type
            -Method and Description
            -
            -
            -ExtendedCell
            -deepClone()
            -Does a deep copy of the contents to a new memory area and 
            returns it as a new cell.
            -
            -
            -
            -int
            -getSerializedSize(booleanwithTags)
            -
            -
            -byte[]
            -getValueArray()
            -Contiguous raw bytes that may start at any index in the 
            containing array.
            -
            -
            -
            -http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
             title="class or interface in java.nio">ByteBuffer
            -getValueByteBuffer()
            -
            -
            -int
            -getValueLength()
            -
            -
            -int
            -getValueOffset()
            -
            -
            -int
            -getValuePosition()
            -
            -
            -long
            -heapSize()
            -
            -
            -void
            -write(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
             title="class or interface in java.nio">ByteBufferbuf,
            - intoffset)
            -Write this Cell into the given buf's offset in a KeyValue format.
            -
            -
            -
            -int
            -write(http://docs.oracle.com/javase/8/docs/api/java/io/OutputStream.html?is-external=true;
             title="class or interface in java.io">OutputStreamout,
            - booleanwithTags)
            -Write this cell to an OutputStream in a KeyValue format.
            -
            -
            -
            -
            -
            -
            -
            -Methods inherited from classorg.apache.hadoop.hbase.PrivateCellUtil.TagRewriteByteBufferCell
            -getFamilyArray,
             getFamilyByteBuffer,
             getFamilyLength,
             getFamilyOffset,
             getFamilyPosition,
             getQualifierArray,
             getQualifierByteBuffer,
             getQualifierLength,
             getQualifierOffset,
             getQualifierPosition,
             getRowArray,
             getRowByteBuffer,
             getRowLength,
             getRowOffset,
             

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OffheapDecodedCell.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OffheapDecodedCell.html
             
            b/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OffheapDecodedCell.html
            index a2c58d5..41b74b8 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OffheapDecodedCell.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OffheapDecodedCell.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10};
            +var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10};
             var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -protected static class BufferedDataBlockEncoder.OffheapDecodedCell
            +protected static class BufferedDataBlockEncoder.OffheapDecodedCell
             extends ByteBufferCell
             implements ExtendedCell
             
            @@ -368,101 +368,89 @@ implements getSerializedSize(booleanwithTags)
             
             
            -http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
             title="class or interface in java.util">OptionalTag
            -getTag(bytetype)
            -Returns the specific tag of the given type
            -
            -
            -
            -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in java.util">ListTag
            -getTags()
            -Creates a list of tags in the current cell
            -
            -
            -
             byte[]
             getTagsArray()
             Contiguous raw bytes representing tags that may start at 
            any index in the containing array.
             
             
            -
            +
             http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
             title="class or interface in java.nio">ByteBuffer
             getTagsByteBuffer()
             
            -
            +
             int
             getTagsLength()
             HBase internally uses 2 bytes to store tags length in 
            Cell.
             
             
            -
            +
             int
             getTagsOffset()
             
            -
            +
             int
             getTagsPosition()
             
            -
            +
             long
             getTimestamp()
             
            -
            +
             byte
             getTypeByte()
             
            -
            +
             byte[]
             getValueArray()
             Contiguous raw bytes that may start at any index in the 
            containing array.
             
             
            -
            +
             http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
             title="class or interface in java.nio">ByteBuffer
             getValueByteBuffer()
             
            -
            +
             int
             getValueLength()
             
            -
            +
             int
             getValueOffset()
             
            -
            +
             int
             getValuePosition()
             
            -
            +
             long
             heapSize()
             
            -
            +
             void
             setSequenceId(longseqId)
             Sets with the given seqId.
             
             
            -
            +
             void
             setTimestamp(byte[]ts)
             Sets with the given timestamp.
             
             
            -
            +
             void
             setTimestamp(longts)
             Sets with the given timestamp.
             
             
            -
            +
             void
             write(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
             title="class or interface in java.nio">ByteBufferbuf,
              intoffset)
             Write this Cell into the given buf's offset in a KeyValue format.
             
             
            -
            +
             int
             write(http://docs.oracle.com/javase/8/docs/api/java/io/OutputStream.html?is-external=true;
             title="class or interface in java.io">OutputStreamout,
              booleanwithTags)
            @@ -489,7 +477,7 @@ implements 
             
             Methods inherited from interfaceorg.apache.hadoop.hbase.RawCell
            -checkForTagsLength,
             cloneTags
            +checkForTagsLength,
             cloneTags,
             getTag,
             getTags
             
             
             
            @@ -511,7 +499,7 @@ implements 
             
             FIXED_OVERHEAD
            -private static finallong FIXED_OVERHEAD
            +private static finallong FIXED_OVERHEAD
             
             
             
            @@ -520,7 +508,7 @@ implements 
             
             keyBuffer
            -privatehttp://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
             title="class or interface in java.nio">ByteBuffer keyBuffer
            +privatehttp://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
             title="class or interface in java.nio">ByteBuffer keyBuffer
             
             
             
            @@ -529,7 +517,7 @@ implements 
             
             rowLength
            -privateshort rowLength
            +privateshort rowLength
             
             
             
            @@ -538,7 +526,7 @@ implements 
             
             familyOffset
            -privateint familyOffset
            +privateint familyOffset
             
             
             
            @@ -547,7 +535,7 @@ implements 
             
             familyLength
            -privatebyte familyLength
            +privatebyte familyLength
             
             
             
            @@ -556,7 +544,7 @@ implements 
             
             qualifierOffset
            -privateint qualifierOffset
            +privateint qualifierOffset
             
             
             
            @@ -565,7 +553,7 @@ implements 
             
             qualifierLength
            -privateint qualifierLength
            +privateint 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
             
            b/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            index 6b833fe..0f2f580 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -static enum MasterRpcServices.BalanceSwitchMode
            +static enum MasterRpcServices.BalanceSwitchMode
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
             title="class or interface in java.lang">EnumMasterRpcServices.BalanceSwitchMode
             
             
            @@ -210,7 +210,7 @@ the order they are declared.
             
             
             SYNC
            -public static finalMasterRpcServices.BalanceSwitchMode SYNC
            +public static finalMasterRpcServices.BalanceSwitchMode SYNC
             
             
             
            @@ -219,7 +219,7 @@ the order they are declared.
             
             
             ASYNC
            -public static finalMasterRpcServices.BalanceSwitchMode ASYNC
            +public static finalMasterRpcServices.BalanceSwitchMode ASYNC
             
             
             
            
            
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html
            index 875ebe0..3bc0e3b 100644
            --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html
            +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html
            @@ -26,724 +26,723 @@
             018package org.apache.hadoop.hbase.client;
             019
             020import com.google.protobuf.RpcChannel;
            -021import java.util.Collection;
            -022import java.util.EnumSet;
            -023import java.util.List;
            -024import java.util.Map;
            -025import java.util.Optional;
            -026import java.util.Set;
            -027import 
            java.util.concurrent.CompletableFuture;
            -028import 
            java.util.concurrent.ExecutorService;
            -029import java.util.function.Function;
            -030import java.util.regex.Pattern;
            -031import 
            org.apache.hadoop.hbase.ClusterMetrics.Option;
            -032import 
            org.apache.hadoop.hbase.ClusterStatus;
            -033import 
            org.apache.hadoop.hbase.NamespaceDescriptor;
            -034import 
            org.apache.hadoop.hbase.RegionLoad;
            -035import 
            org.apache.hadoop.hbase.ServerName;
            -036import 
            org.apache.hadoop.hbase.TableName;
            -037import 
            org.apache.hadoop.hbase.client.replication.TableCFs;
            -038import 
            org.apache.hadoop.hbase.client.security.SecurityCapability;
            -039import 
            org.apache.hadoop.hbase.quotas.QuotaFilter;
            -040import 
            org.apache.hadoop.hbase.quotas.QuotaSettings;
            -041import 
            org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
            -042import 
            org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
            -043import 
            org.apache.yetus.audience.InterfaceAudience;
            -044
            -045/**
            -046 * Just a wrapper of {@link 
            RawAsyncHBaseAdmin}. The difference is that users need to provide a
            -047 * thread pool when constructing this 
            class, and the callback methods registered to the returned
            -048 * {@link CompletableFuture} will be 
            executed in this thread pool. So usually it is safe for users
            -049 * to do anything they want in the 
            callbacks without breaking the rpc framework.
            -050 * @since 2.0.0
            -051 * @see RawAsyncHBaseAdmin
            -052 * @see 
            AsyncConnection#getAdmin(ExecutorService)
            -053 * @see 
            AsyncConnection#getAdminBuilder(ExecutorService)
            -054 */
            -055@InterfaceAudience.Private
            -056class AsyncHBaseAdmin implements 
            AsyncAdmin {
            -057
            -058  private final RawAsyncHBaseAdmin 
            rawAdmin;
            -059
            -060  private final ExecutorService pool;
            -061
            -062  AsyncHBaseAdmin(RawAsyncHBaseAdmin 
            rawAdmin, ExecutorService pool) {
            -063this.rawAdmin = rawAdmin;
            -064this.pool = pool;
            -065  }
            -066
            -067  private T 
            CompletableFutureT wrap(CompletableFutureT future) {
            -068CompletableFutureT 
            asyncFuture = new CompletableFuture();
            -069future.whenCompleteAsync((r, e) - 
            {
            -070  if (e != null) {
            -071
            asyncFuture.completeExceptionally(e);
            -072  } else {
            -073asyncFuture.complete(r);
            -074  }
            -075}, pool);
            -076return asyncFuture;
            -077  }
            -078
            -079  @Override
            -080  public CompletableFutureBoolean 
            tableExists(TableName tableName) {
            -081return 
            wrap(rawAdmin.tableExists(tableName));
            -082  }
            -083
            -084  @Override
            -085  public 
            CompletableFutureListTableDescriptor 
            listTableDescriptors(boolean includeSysTables) {
            -086return 
            wrap(rawAdmin.listTableDescriptors(includeSysTables));
            -087  }
            -088
            -089  @Override
            -090  public 
            CompletableFutureListTableDescriptor 
            listTableDescriptors(Pattern pattern,
            -091  boolean includeSysTables) {
            -092return 
            wrap(rawAdmin.listTableDescriptors(pattern, includeSysTables));
            -093  }
            -094
            -095  @Override
            -096  public 
            CompletableFutureListTableDescriptor 
            listTableDescriptorsByNamespace(String name) {
            -097return 
            wrap(rawAdmin.listTableDescriptorsByNamespace(name));
            -098  }
            -099
            -100  @Override
            -101  public 
            CompletableFutureListTableName listTableNames(boolean 
            includeSysTables) {
            -102return 
            wrap(rawAdmin.listTableNames(includeSysTables));
            -103  }
            -104
            -105  @Override
            -106  public 
            CompletableFutureListTableName listTableNames(Pattern 
            pattern,
            -107  boolean includeSysTables) {
            -108return 
            wrap(rawAdmin.listTableNames(pattern, includeSysTables));
            -109  }
            -110
            -111  @Override
            -112  public 
            CompletableFutureListTableName listTableNamesByNamespace(String 
            name) {
            -113return 
            wrap(rawAdmin.listTableNamesByNamespace(name));
            -114  }
            -115
            -116  @Override
            -117  public 
            CompletableFutureTableDescriptor getDescriptor(TableName tableName) {
            -118return 
            wrap(rawAdmin.getDescriptor(tableName));
            -119  }
            -120
            -121  @Override
            -122  public CompletableFutureVoid 
            createTable(TableDescriptor desc) {
            -123return 
            wrap(rawAdmin.createTable(desc));
            -124  }
            -125
            -126  @Override
            -127  public CompletableFutureVoid 
            createTable(TableDescriptor desc, byte[] startKey, byte[] endKey,
            -128  int numRegions) {
            -129return 
            wrap(rawAdmin.createTable(desc, startKey, 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/apidocs/src-html/org/apache/hadoop/hbase/client/Append.html
            --
            diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Append.html 
            b/apidocs/src-html/org/apache/hadoop/hbase/client/Append.html
            index bed985e..809287d 100644
            --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Append.html
            +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Append.html
            @@ -120,131 +120,143 @@
             112  }
             113  /**
             114   * Copy constructor
            -115   * @param a
            +115   * @param appendToCopy append to copy
             116   */
            -117  public Append(Append a) {
            -118this.row = a.getRow();
            -119this.ts = a.getTimeStamp();
            -120this.tr = a.getTimeRange();
            -121
            this.familyMap.putAll(a.getFamilyCellMap());
            -122for (Map.EntryString, byte[] 
            entry : a.getAttributesMap().entrySet()) {
            -123  this.setAttribute(entry.getKey(), 
            entry.getValue());
            -124}
            -125this.setPriority(a.getPriority());
            -126  }
            -127
            -128  /** Create a Append operation for the 
            specified row.
            -129   * p
            -130   * At least one column must be appended 
            to.
            -131   * @param rowArray Makes a copy out of 
            this buffer.
            -132   * @param rowOffset
            -133   * @param rowLength
            -134   */
            -135  public Append(final byte [] rowArray, 
            final int rowOffset, final int rowLength) {
            -136checkRow(rowArray, rowOffset, 
            rowLength);
            -137this.row = Bytes.copy(rowArray, 
            rowOffset, rowLength);
            -138  }
            -139
            -140  /**
            -141   * Add the specified column and value 
            to this Append operation.
            -142   * @param family family name
            -143   * @param qualifier column qualifier
            -144   * @param value value to append to 
            specified column
            -145   * @return this
            -146   * @deprecated As of release 2.0.0, 
            this will be removed in HBase 3.0.0.
            -147   * Use {@link 
            #addColumn(byte[], byte[], byte[])} instead
            -148   */
            -149  @Deprecated
            -150  public Append add(byte [] family, byte 
            [] qualifier, byte [] value) {
            -151return this.addColumn(family, 
            qualifier, value);
            -152  }
            -153
            -154  /**
            -155   * Add the specified column and value 
            to this Append operation.
            -156   * @param family family name
            -157   * @param qualifier column qualifier
            -158   * @param value value to append to 
            specified column
            -159   * @return this
            -160   */
            -161  public Append addColumn(byte[] family, 
            byte[] qualifier, byte[] value) {
            -162KeyValue kv = new KeyValue(this.row, 
            family, qualifier, this.ts, KeyValue.Type.Put, value);
            -163return add(kv);
            -164  }
            -165
            -166  /**
            -167   * Add column and value to this Append 
            operation.
            -168   * @param cell
            -169   * @return This instance
            -170   */
            -171  @SuppressWarnings("unchecked")
            -172  public Append add(final Cell cell) {
            -173// Presume it is KeyValue for now.
            -174byte [] family = 
            CellUtil.cloneFamily(cell);
            -175
            -176// Get cell list for the family
            -177ListCell list = 
            getCellList(family);
            -178
            -179// find where the new entry should be 
            placed in the List
            -180list.add(cell);
            -181return this;
            -182  }
            -183
            -184  @Override
            -185  public Append setTimestamp(long 
            timestamp) {
            -186super.setTimestamp(timestamp);
            +117  public Append(Append appendToCopy) {
            +118super(appendToCopy);
            +119this.tr = 
            appendToCopy.getTimeRange();
            +120  }
            +121
            +122  /** Create a Append operation for the 
            specified row.
            +123   * p
            +124   * At least one column must be appended 
            to.
            +125   * @param rowArray Makes a copy out of 
            this buffer.
            +126   * @param rowOffset
            +127   * @param rowLength
            +128   */
            +129  public Append(final byte [] rowArray, 
            final int rowOffset, final int rowLength) {
            +130checkRow(rowArray, rowOffset, 
            rowLength);
            +131this.row = Bytes.copy(rowArray, 
            rowOffset, rowLength);
            +132  }
            +133
            +134  /**
            +135   * Construct the Append with user 
            defined data. NOTED:
            +136   * 1) all cells in the familyMap must 
            have the DataType.Put
            +137   * 2) the row of each cell must be same 
            with passed row.
            +138   * @param row row. CAN'T be null
            +139   * @param ts timestamp
            +140   * @param familyMap the map to collect 
            all cells internally. CAN'T be null
            +141   */
            +142  public Append(byte[] row, long ts, 
            NavigableMapbyte [], ListCell familyMap) {
            +143super(row, ts, familyMap);
            +144  }
            +145
            +146  /**
            +147   * Add the specified column and value 
            to this Append operation.
            +148   * @param family family name
            +149   * @param qualifier column qualifier
            +150   * @param value value to append to 
            specified column
            +151   * @return this
            +152   * @deprecated As of release 2.0.0, 
            this will be removed in HBase 3.0.0.
            +153   * Use {@link 
            #addColumn(byte[], byte[], byte[])} instead
            +154   */
            +155  @Deprecated
            +156  public Append add(byte [] family, byte 
            [] qualifier, byte [] value) {
            +157return this.addColumn(family, 
            qualifier, value);
            +158  }
            +159
            +160  /**
            +161   * Add the specified column and value 
            to this Append operation.
            +162   * @param family family name
            +163   * @param qualifier column qualifier
            +164   * 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64InputStream.html
            --
            diff --git 
            a/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64InputStream.html 
            b/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64InputStream.html
            index 70481ce..b9f6622 100644
            --- 
            a/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64InputStream.html
            +++ 
            b/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64InputStream.html
            @@ -47,1648 +47,1649 @@
             039import java.util.zip.GZIPInputStream;
             040import java.util.zip.GZIPOutputStream;
             041
            -042import org.apache.commons.logging.Log;
            -043import 
            org.apache.commons.logging.LogFactory;
            -044import 
            org.apache.yetus.audience.InterfaceAudience;
            -045
            -046/**
            -047 * Encodes and decodes to and from Base64 
            notation.
            -048 *
            -049 * p
            -050 * Homepage: a 
            href="http://iharder.net/base64"http://iharder.net/base64/a;.
            -051 * /p
            -052 *
            -053 * p
            -054 * Change Log:
            -055 * /p
            -056 * ul
            -057 *   liv2.2.1 - Fixed bug using 
            URL_SAFE and ORDERED encodings. Fixed bug
            -058 * when using very small files 
            (~lt; 40 bytes)./li
            -059 *   liv2.2 - Added some helper 
            methods for encoding/decoding directly from
            -060 * one file to the next. Also added a 
            main() method to support command
            -061 * line encoding/decoding from one 
            file to the next. Also added these
            -062 * Base64 dialects:
            -063 * ol
            -064 *   liThe default is RFC3548 
            format./li
            -065 *   liUsing Base64.URLSAFE 
            generates URL and file name friendly format as
            -066 * described in Section 4 of 
            RFC3548.
            -067 * 
            http://www.faqs.org/rfcs/rfc3548.html/li;
            -068 *   liUsing Base64.ORDERED 
            generates URL and file name friendly format
            -069 * that preserves lexical 
            ordering as described in
            -070 * 
            http://www.faqs.org/qa/rfcc-1940.html/li;
            -071 * /ol
            -072 * p
            -073 * Special thanks to Jim Kellerman at 
            a href="http://www.powerset.com/";
            -074 * http://www.powerset.com//a; 
            for contributing the new Base64 dialects.
            -075 *   /li
            -076 *
            -077 *   liv2.1 - Cleaned up javadoc 
            comments and unused variables and methods.
            -078 * Added some convenience methods for 
            reading and writing to and from files.
            -079 *   /li
            -080 *   liv2.0.2 - Now specifies 
            UTF-8 encoding in places where the code fails on
            -081 * systems with other encodings (like 
            EBCDIC)./li
            -082 *   liv2.0.1 - Fixed an error 
            when decoding a single byte, that is, when the
            -083 * encoded data was a single 
            byte./li
            -084 *   liv2.0 - I got rid of 
            methods that used booleans to set options. Now
            -085 * everything is more consolidated 
            and cleaner. The code now detects when
            -086 * data that's being decoded is 
            gzip-compressed and will decompress it
            -087 * automatically. Generally things 
            are cleaner. You'll probably have to
            -088 * change some method calls that you 
            were making to support the new options
            -089 * format (ttint/tts 
            that you "OR" together)./li
            -090 *   liv1.5.1 - Fixed bug when 
            decompressing and decoding to a byte[] using
            -091 * ttdecode( String s, 
            boolean gzipCompressed )/tt. Added the ability to
            -092 * "suspend" encoding in the Output 
            Stream so you can turn on and off the
            -093 * encoding if you need to embed 
            base64 data in an otherwise "normal" stream
            -094 * (like an XML file)./li
            -095 *   liv1.5 - Output stream pases 
            on flush() command but doesn't do anything
            -096 * itself. This helps when using GZIP 
            streams. Added the ability to
            -097 * GZip-compress objects before 
            encoding them./li
            -098 *   liv1.4 - Added helper 
            methods to read/write files./li
            -099 *   liv1.3.6 - Fixed 
            OutputStream.flush() so that 'position' is reset./li
            -100 *   liv1.3.5 - Added flag to 
            turn on and off line breaks. Fixed bug in input
            -101 * stream where last buffer being 
            read, if not completely full, was not
            -102 * returned./li
            -103 *   liv1.3.4 - Fixed when 
            "improperly padded stream" error was thrown at the
            -104 * wrong time./li
            -105 *   liv1.3.3 - Fixed I/O streams 
            which were totally messed up./li
            -106 * /ul
            -107 *
            -108 * p
            -109 * I am placing this code in the Public 
            Domain. Do with it as you will. This
            -110 * software comes with no guarantees or 
            warranties but with plenty of
            -111 * well-wishing instead!
            -112 * p
            -113 * Please visit a 
            href="http://iharder.net/base64"http://iharder.net/base64/a;
            -114 * periodically to check for updates or 
            to contribute improvements.
            -115 * p
            -116 * author: Robert Harder, 
            r...@iharder.net
            -117 * br
            -118 * version: 2.2.1
            -119 */
            -120@InterfaceAudience.Public
            -121public class Base64 {
            -122
            -123  /*  P U B L I C   F I E L D S 
             */
            -124
            -125  /** No options specified. Value is 
            zero. */
            -126  public final static int NO_OPTIONS = 
            0;
            -127
            -128  /** Specify encoding. */
            -129  public final static int ENCODE = 1;
            -130
            -131  /** Specify decoding. */
            -132  public final 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/filter/BitComparator.BitwiseOp.html
            --
            diff --git 
            a/apidocs/src-html/org/apache/hadoop/hbase/filter/BitComparator.BitwiseOp.html 
            b/apidocs/src-html/org/apache/hadoop/hbase/filter/BitComparator.BitwiseOp.html
            index ae3605c..3179892 100644
            --- 
            a/apidocs/src-html/org/apache/hadoop/hbase/filter/BitComparator.BitwiseOp.html
            +++ 
            b/apidocs/src-html/org/apache/hadoop/hbase/filter/BitComparator.BitwiseOp.html
            @@ -75,95 +75,97 @@
             067  /**
             068   * @return The comparator serialized 
            using pb
             069   */
            -070  public byte [] toByteArray() {
            -071
            ComparatorProtos.BitComparator.Builder builder =
            -072  
            ComparatorProtos.BitComparator.newBuilder();
            -073
            builder.setComparable(ProtobufUtil.toByteArrayComparable(this.value));
            -074
            ComparatorProtos.BitComparator.BitwiseOp bitwiseOpPb =
            -075  
            ComparatorProtos.BitComparator.BitwiseOp.valueOf(bitOperator.name());
            -076builder.setBitwiseOp(bitwiseOpPb);
            -077return 
            builder.build().toByteArray();
            -078  }
            -079
            -080  /**
            -081   * @param pbBytes A pb serialized 
            {@link BitComparator} instance
            -082   * @return An instance of {@link 
            BitComparator} made from codebytes/code
            -083   * @throws DeserializationException
            -084   * @see #toByteArray
            -085   */
            -086  public static BitComparator 
            parseFrom(final byte [] pbBytes)
            -087  throws DeserializationException {
            -088ComparatorProtos.BitComparator 
            proto;
            -089try {
            -090  proto = 
            ComparatorProtos.BitComparator.parseFrom(pbBytes);
            -091} catch 
            (InvalidProtocolBufferException e) {
            -092  throw new 
            DeserializationException(e);
            -093}
            -094BitwiseOp bitwiseOp = 
            BitwiseOp.valueOf(proto.getBitwiseOp().name());
            -095return new 
            BitComparator(proto.getComparable().getValue().toByteArray(),bitwiseOp);
            -096  }
            -097
            -098  /**
            -099   * @param other
            -100   * @return true if and only if the 
            fields of the comparator that are serialized
            -101   * are equal to the corresponding 
            fields in other.  Used for testing.
            -102   */
            -103  boolean 
            areSerializedFieldsEqual(ByteArrayComparable other) {
            -104if (other == this) return true;
            -105if (!(other instanceof 
            BitComparator)) return false;
            -106
            -107BitComparator comparator = 
            (BitComparator)other;
            -108return 
            super.areSerializedFieldsEqual(other)
            -109   
            this.getOperator().equals(comparator.getOperator());
            -110  }
            -111
            -112  @Override
            -113  public int compareTo(byte[] value, int 
            offset, int length) {
            -114if (length != this.value.length) {
            -115  return 1;
            -116}
            -117int b = 0;
            -118//Iterating backwards is faster 
            because we can quit after one non-zero byte.
            -119for (int i = length - 1; i = 0 
             b == 0; i--) {
            -120  switch (bitOperator) {
            -121case AND:
            -122  b = (this.value[i]  
            value[i+offset])  0xff;
            -123  break;
            -124case OR:
            -125  b = (this.value[i] | 
            value[i+offset])  0xff;
            -126  break;
            -127case XOR:
            -128  b = (this.value[i] ^ 
            value[i+offset])  0xff;
            -129  break;
            -130  }
            -131}
            -132return b == 0 ? 1 : 0;
            -133  }
            -134
            -135  @Override
            -136  public int compareTo(ByteBuffer value, 
            int offset, int length) {
            -137if (length != this.value.length) {
            -138  return 1;
            -139}
            -140int b = 0;
            -141//Iterating backwards is faster 
            because we can quit after one non-zero byte.
            -142for (int i = length - 1; i = 0 
             b == 0; i--) {
            -143  switch (bitOperator) {
            -144case AND:
            -145  b = (this.value[i]  
            value.get(i + offset))  0xff;
            -146  break;
            -147case OR:
            -148  b = (this.value[i] | 
            value.get(i + offset))  0xff;
            -149  break;
            -150case XOR:
            -151  b = (this.value[i] ^ 
            value.get(i + offset))  0xff;
            -152  break;
            -153  }
            -154}
            -155return b == 0 ? 1 : 0;
            -156  }
            -157}
            -158
            +070  @Override
            +071  public byte [] toByteArray() {
            +072
            ComparatorProtos.BitComparator.Builder builder =
            +073  
            ComparatorProtos.BitComparator.newBuilder();
            +074
            builder.setComparable(ProtobufUtil.toByteArrayComparable(this.value));
            +075
            ComparatorProtos.BitComparator.BitwiseOp bitwiseOpPb =
            +076  
            ComparatorProtos.BitComparator.BitwiseOp.valueOf(bitOperator.name());
            +077builder.setBitwiseOp(bitwiseOpPb);
            +078return 
            builder.build().toByteArray();
            +079  }
            +080
            +081  /**
            +082   * @param pbBytes A pb serialized 
            {@link BitComparator} instance
            +083   * @return An instance of {@link 
            BitComparator} made from codebytes/code
            +084   * @throws DeserializationException
            +085   * @see #toByteArray
            +086   */
            +087  public static BitComparator 
            parseFrom(final byte [] pbBytes)
            +088  throws DeserializationException {
            +089ComparatorProtos.BitComparator 
            proto;
            +090try {
            +091  proto = 
            ComparatorProtos.BitComparator.parseFrom(pbBytes);
            +092} catch 
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
             
            b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
            index 16ff501..067d702 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
            @@ -137,7 +137,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -public static class ThriftServerRunner.IOErrorWithCause
            +public static class ThriftServerRunner.IOErrorWithCause
             extends org.apache.hadoop.hbase.thrift.generated.IOError
             
             See Also:
            @@ -273,7 +273,7 @@ extends 
            org.apache.hadoop.hbase.thrift.generated.IOError
             
             
             cause
            -privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
             title="class or interface in java.lang">Throwable cause
            +privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
             title="class or interface in java.lang">Throwable cause
             
             
             
            @@ -290,7 +290,7 @@ extends 
            org.apache.hadoop.hbase.thrift.generated.IOError
             
             
             IOErrorWithCause
            -publicIOErrorWithCause(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
             title="class or interface in java.lang">Throwablecause)
            +publicIOErrorWithCause(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
             title="class or interface in java.lang">Throwablecause)
             
             
             
            @@ -307,7 +307,7 @@ extends 
            org.apache.hadoop.hbase.thrift.generated.IOError
             
             
             getCause
            -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
             title="class or interface in java.lang">ThrowablegetCause()
            +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
             title="class or interface in java.lang">ThrowablegetCause()
             
             Overrides:
             http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getCause--;
             title="class or interface in java.lang">getCausein 
            classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
             title="class or interface in java.lang">Throwable
            @@ -320,7 +320,7 @@ extends 
            org.apache.hadoop.hbase.thrift.generated.IOError
             
             
             equals
            -publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Objectother)
            +publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Objectother)
             
             Overrides:
             equalsin 
            classorg.apache.hadoop.hbase.thrift.generated.IOError
            @@ -333,7 +333,7 @@ extends 
            org.apache.hadoop.hbase.thrift.generated.IOError
             
             
             hashCode
            -publicinthashCode()
            +publicinthashCode()
             
             Overrides:
             hashCodein 
            classorg.apache.hadoop.hbase.thrift.generated.IOError
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html 
            b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
            index 5e08f1c..cd86385 100644
            --- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
            +++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
            @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -static enum ThriftServerRunner.ImplType
            +static enum ThriftServerRunner.ImplType
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
             title="class or interface in java.lang">EnumThriftServerRunner.ImplType
             An enum of server implementation selections
             
            @@ -281,7 +281,7 @@ the order they are declared.
             
             
             HS_HA
            -public static finalThriftServerRunner.ImplType 
            HS_HA
            +public static finalThriftServerRunner.ImplType 
            HS_HA
             
             
             
            @@ -290,7 +290,7 @@ the order they are declared.
             
             
             NONBLOCKING
            -public static finalThriftServerRunner.ImplType 
            NONBLOCKING
            +public static finalThriftServerRunner.ImplType 
            NONBLOCKING
             
             
             
            @@ -299,7 +299,7 @@ the order they are declared.
             
             
             THREAD_POOL
            -public static finalThriftServerRunner.ImplType 
            THREAD_POOL
            +public static finalThriftServerRunner.ImplType 
            THREAD_POOL
             
             
             
            @@ -308,7 +308,7 @@ the order they are declared.
             
             
             THREADED_SELECTOR
            -public static finalThriftServerRunner.ImplType 
            THREADED_SELECTOR
            +public static finalThriftServerRunner.ImplType 
            THREADED_SELECTOR
             
             
             
            @@ -325,7 +325,7 @@ the order they are declared.
             
             
             DEFAULT
            -public static finalThriftServerRunner.ImplType 
            DEFAULT
            +public static finalThriftServerRunner.ImplType 
            DEFAULT
             
             
             
            @@ -334,7 +334,7 @@ the order 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteTableFuture.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteTableFuture.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteTableFuture.html
            index 7c59e27..c904c56 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteTableFuture.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteTableFuture.html
            @@ -119,4048 +119,4054 @@
             111import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
             112import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
             113import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearRegionBlockCacheRequest;
            -114import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
            -115import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
            -116import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
            -117import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
            -118import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest;
            -119import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse;
            -120import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest;
            -121import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest;
            -122import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
            -123import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest;
            -124import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse;
            -125import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
            -126import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
            -127import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
            -128import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
            -129import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos;
            -130import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
            -131import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
            -132import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
            -133import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
            -134import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
            -135import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest;
            -136import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
            -137import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
            -138import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest;
            -139import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse;
            -140import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
            -141import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
            -142import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
            -143import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse;
            -144import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest;
            -145import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest;
            -146import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse;
            -147import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
            -148import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
            -149import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
            -150import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
            -151import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest;
            -152import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse;
            -153import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest;
            -154import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest;
            -155import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest;
            -156import 
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html 
            b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
            index 1cf9ab3..767aa3b 100644
            --- a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
            +++ b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
            @@ -1583,7 +1583,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
             
             
             createPutForStartBackupSession
            -privatePutcreatePutForStartBackupSession()
            +privatePutcreatePutForStartBackupSession()
             
             
             
            @@ -1592,7 +1592,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
             
             
             finishBackupExclusiveOperation
            -publicvoidfinishBackupExclusiveOperation()
            +publicvoidfinishBackupExclusiveOperation()
             throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
             title="class or interface in java.io">IOException
             
             Throws:
            @@ -1606,7 +1606,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
             
             
             createPutForStopBackupSession
            -privatePutcreatePutForStopBackupSession()
            +privatePutcreatePutForStopBackupSession()
             
             
             
            @@ -1615,7 +1615,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
             
             
             readRegionServerLastLogRollResult
            -publichttp://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true;
             title="class or interface in java.util">HashMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
             title="class or interface in java.lang">LongreadRegionServerLastLogRollResult(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringbackupRoot)
            +publichttp://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true;
             title="class or interface in java.util">HashMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
             title="class or interface in java.lang">LongreadRegionServerLastLogRollResult(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringbackupRoot)
            throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
             title="class or interface in java.io">IOException
             Get the Region Servers log information after the last log 
            roll from backup system table.
             
            @@ -1634,7 +1634,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
             
             
             writeRegionServerLastLogRollResult
            -publicvoidwriteRegionServerLastLogRollResult(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringserver,
            +publicvoidwriteRegionServerLastLogRollResult(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringserver,
            http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
             title="class or interface in java.lang">Longts,
            http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">StringbackupRoot)
             throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
             title="class or interface in java.io">IOException
            @@ -1655,7 +1655,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
             
             
             getBackupHistory
            -publichttp://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true;
             title="class or interface in java.util">ArrayListBackupInfogetBackupHistory(booleanonlyCompleted)
            +publichttp://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true;
             title="class or interface in java.util">ArrayListBackupInfogetBackupHistory(booleanonlyCompleted)
            throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
             title="class or interface in java.io">IOException
             Get all completed backup information (in desc order by 
            time)
             
            @@ -1674,7 +1674,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
             
             
             getBackupHistory
            -publichttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
             title="class or interface in 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/IndividualBytesFieldCell.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/IndividualBytesFieldCell.html 
            b/devapidocs/org/apache/hadoop/hbase/IndividualBytesFieldCell.html
            index 06d9a30..c95c548 100644
            --- a/devapidocs/org/apache/hadoop/hbase/IndividualBytesFieldCell.html
            +++ b/devapidocs/org/apache/hadoop/hbase/IndividualBytesFieldCell.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10};
            +var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10};
             var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
            Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -109,12 +109,12 @@ var activeTableTab = "activeTableTab";
             
             
             All Implemented Interfaces:
            -http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true;
             title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell, SettableSequenceId, SettableTimestamp
            +http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true;
             title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell
             
             
             
             @InterfaceAudience.Private
            -public class IndividualBytesFieldCell
            +public class IndividualBytesFieldCell
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
             implements ExtendedCell
             
            @@ -305,149 +305,124 @@ implements 
             
             
            -ExtendedCell
            -deepClone()
            -Does a deep copy of the contents to a new memory area and 
            returns it as a new cell.
            -
            -
            -
             byte[]
             getFamilyArray()
             Contiguous bytes composed of legal HDFS filename characters 
            which may start at any index in the
              containing array.
             
             
            -
            +
             byte
             getFamilyLength()
             
            -
            +
             int
             getFamilyOffset()
             
            -
            +
             byte[]
             getQualifierArray()
             Contiguous raw bytes that may start at any index in the 
            containing array.
             
             
            -
            +
             int
             getQualifierLength()
             
            -
            +
             int
             getQualifierOffset()
             
            -
            +
             byte[]
             getRowArray()
             Implement Cell interface
             
             
            -
            +
             short
             getRowLength()
             
            -
            +
             int
             getRowOffset()
             
            -
            +
             long
             getSequenceId()
             A region-specific unique monotonically increasing sequence 
            ID given to each Cell.
             
             
            -
            -int
            -getSerializedSize(booleanwithTags)
            -
            -
            +
             byte[]
             getTagsArray()
             Contiguous raw bytes representing tags that may start at 
            any index in the containing array.
             
             
            -
            +
             int
             getTagsLength()
             HBase internally uses 2 bytes to store tags length in 
            Cell.
             
             
            -
            +
             int
             getTagsOffset()
             
            -
            +
             long
             getTimestamp()
             
            -
            +
             byte
             getTypeByte()
             
            -
            +
             byte[]
             getValueArray()
             Contiguous raw bytes that may start at any index in the 
            containing array.
             
             
            -
            +
             int
             getValueLength()
             
            -
            +
             int
             getValueOffset()
             
            -
            +
             private long
             heapOverhead()
             
            -
            +
             long
             heapSize()
             Implement HeapSize interface
             
             
            -
            +
             void
             setSequenceId(longseqId)
            -Implement SettableSequenceId interface
            +Sets with the given seqId.
             
             
            -
            +
             void
            -setTimestamp(byte[]ts,
            -inttsOffset)
            +setTimestamp(byte[]ts)
             Sets with the given timestamp.
             
             
            -
            +
             void
             setTimestamp(longts)
            -Implement SettableTimestamp interface
            +Sets with the given timestamp.
             
             
            -
            +
             http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">String
             toString()
             
            -
            -void
            -write(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
             title="class or interface in java.nio">ByteBufferbuf,
            - intoffset)
            -Write this Cell into the given buf's offset in a KeyValue format.
            -
            -
            -
            -int
            -write(http://docs.oracle.com/javase/8/docs/api/java/io/OutputStream.html?is-external=true;
             title="class or interface in java.io">OutputStreamout,
            - booleanwithTags)
            -Write this cell to an OutputStream in a KeyValue format.
            -
            -
             
             
             
            @@ -461,7 +436,7 @@ implements 
             
             Methods inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell
            -getChunkId
            +deepClone,
             getChunkId,
             getSerializedSize,
             write,
             write
             
             
             
            @@ -490,7 +465,7 @@ implements 
             
             FIXED_OVERHEAD
            -private static finallong FIXED_OVERHEAD
            +private static finallong FIXED_OVERHEAD
             
             
             
            @@ -499,7 +474,7 @@ implements 
             
             row
            -private finalbyte[] row
            +private finalbyte[] row
             
             
             
            @@ -508,7 +483,7 @@ implements 
             
             rOffset
            -private finalint rOffset
            +private finalint rOffset
             
             
             
            @@ -517,7 +492,7 @@ implements 
             
             rLength
            -private finalint rLength
            +private finalint rLength
             
             
             
            @@ -526,7 +501,7 @@ 

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/src-html/org/apache/hadoop/hbase/client/Put.html
            --
            diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/Put.html 
            b/devapidocs/src-html/org/apache/hadoop/hbase/client/Put.html
            index 7d3d133..75a155d 100644
            --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/Put.html
            +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/Put.html
            @@ -35,487 +35,355 @@
             027import java.util.NavigableMap;
             028import java.util.TreeMap;
             029import java.util.UUID;
            -030
            -031import org.apache.hadoop.hbase.Cell;
            -032import 
            org.apache.hadoop.hbase.CellUtil;
            -033import 
            org.apache.hadoop.hbase.HConstants;
            +030import org.apache.hadoop.hbase.Cell;
            +031import 
            org.apache.hadoop.hbase.CellUtil;
            +032import 
            org.apache.hadoop.hbase.HConstants;
            +033import 
            org.apache.hadoop.hbase.IndividualBytesFieldCell;
             034import 
            org.apache.hadoop.hbase.KeyValue;
            -035import 
            org.apache.hadoop.hbase.IndividualBytesFieldCell;
            -036import org.apache.hadoop.hbase.Tag;
            -037import 
            org.apache.yetus.audience.InterfaceAudience;
            -038import 
            org.apache.hadoop.hbase.io.HeapSize;
            -039import 
            org.apache.hadoop.hbase.security.access.Permission;
            -040import 
            org.apache.hadoop.hbase.security.visibility.CellVisibility;
            -041import 
            org.apache.hadoop.hbase.util.Bytes;
            -042
            -043/**
            -044 * Used to perform Put operations for a 
            single row.
            -045 * p
            -046 * To perform a Put, instantiate a Put 
            object with the row to insert to, and
            -047 * for each column to be inserted, 
            execute {@link #addColumn(byte[], byte[],
            -048 * byte[]) add} or {@link 
            #addColumn(byte[], byte[], long, byte[]) add} if
            -049 * setting the timestamp.
            -050 */
            -051@InterfaceAudience.Public
            -052public class Put extends Mutation 
            implements HeapSize, ComparableRow {
            -053  /**
            -054   * Create a Put operation for the 
            specified row.
            -055   * @param row row key
            -056   */
            -057  public Put(byte [] row) {
            -058this(row, 
            HConstants.LATEST_TIMESTAMP);
            -059  }
            -060
            -061  /**
            -062   * Create a Put operation for the 
            specified row, using a given timestamp.
            -063   *
            -064   * @param row row key; we make a copy 
            of what we are passed to keep local.
            -065   * @param ts timestamp
            -066   */
            -067  public Put(byte[] row, long ts) {
            -068this(row, 0, row.length, ts);
            -069  }
            -070
            -071  /**
            -072   * We make a copy of the passed in row 
            key to keep local.
            -073   * @param rowArray
            -074   * @param rowOffset
            -075   * @param rowLength
            -076   */
            -077  public Put(byte [] rowArray, int 
            rowOffset, int rowLength) {
            -078this(rowArray, rowOffset, rowLength, 
            HConstants.LATEST_TIMESTAMP);
            -079  }
            -080
            -081  /**
            -082   * @param row row key; we make a copy 
            of what we are passed to keep local.
            -083   * @param ts  timestamp
            -084   */
            -085  public Put(ByteBuffer row, long ts) {
            -086if (ts  0) {
            -087  throw new 
            IllegalArgumentException("Timestamp cannot be negative. ts=" + ts);
            -088}
            -089checkRow(row);
            -090this.row = new 
            byte[row.remaining()];
            -091row.get(this.row);
            -092this.ts = ts;
            -093  }
            -094
            -095  /**
            -096   * @param row row key; we make a copy 
            of what we are passed to keep local.
            -097   */
            -098  public Put(ByteBuffer row) {
            -099this(row, 
            HConstants.LATEST_TIMESTAMP);
            -100  }
            -101
            -102  /**
            -103   * We make a copy of the passed in row 
            key to keep local.
            -104   * @param rowArray
            -105   * @param rowOffset
            -106   * @param rowLength
            -107   * @param ts
            -108   */
            -109  public Put(byte [] rowArray, int 
            rowOffset, int rowLength, long ts) {
            -110checkRow(rowArray, rowOffset, 
            rowLength);
            -111this.row = Bytes.copy(rowArray, 
            rowOffset, rowLength);
            -112this.ts = ts;
            -113if (ts  0) {
            -114  throw new 
            IllegalArgumentException("Timestamp cannot be negative. ts=" + ts);
            -115}
            -116  }
            -117
            -118  /**
            -119   * Create a Put operation for an 
            immutable row key.
            -120   *
            -121   * @param row row key
            -122   * @param rowIsImmutable whether the 
            input row is immutable.
            -123   *   Set to true if 
            the caller can guarantee that
            -124   *   the row will 
            not be changed for the Put duration.
            -125   */
            -126  public Put(byte [] row, boolean 
            rowIsImmutable) {
            -127this(row, 
            HConstants.LATEST_TIMESTAMP, rowIsImmutable);
            -128  }
            -129
            -130  /**
            -131   * Create a Put operation for an 
            immutable row key, using a given timestamp.
            -132   *
            -133   * @param row row key
            -134   * @param ts timestamp
            -135   * @param rowIsImmutable whether the 
            input row is immutable.
            -136   *   Set to true if 
            the caller can guarantee that
            -137   *   the row will 
            not be changed for the Put duration.
            -138   */
            -139  public Put(byte[] row, long ts, boolean 
            rowIsImmutable) {
            -140// Check and set timestamp
            -141if (ts  0) {
            -142  throw new 
            IllegalArgumentException("Timestamp cannot be negative. ts=" + ts);
            -143}
            -144this.ts = ts;
            -145
            -146// Deal with row according to 
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
            index 3edfbef..9707b2c 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
            @@ -2459,5936 +2459,5935 @@
             2451  }
             2452
             2453  for (HStore s : storesToFlush) {
            -2454MemStoreSize flushableSize = 
            s.getFlushableSize();
            -2455
            totalSizeOfFlushableStores.incMemStoreSize(flushableSize);
            -2456
            storeFlushCtxs.put(s.getColumnFamilyDescriptor().getName(),
            -2457  
            s.createFlushContext(flushOpSeqId, tracker));
            -2458// for writing stores to WAL
            -2459
            committedFiles.put(s.getColumnFamilyDescriptor().getName(), null);
            -2460
            storeFlushableSize.put(s.getColumnFamilyDescriptor().getName(), 
            flushableSize);
            -2461  }
            -2462
            -2463  // write the snapshot start to 
            WAL
            -2464  if (wal != null  
            !writestate.readOnly) {
            -2465FlushDescriptor desc = 
            ProtobufUtil.toFlushDescriptor(FlushAction.START_FLUSH,
            -2466getRegionInfo(), 
            flushOpSeqId, committedFiles);
            -2467// No sync. Sync is below where 
            no updates lock and we do FlushAction.COMMIT_FLUSH
            -2468WALUtil.writeFlushMarker(wal, 
            this.getReplicationScope(), getRegionInfo(), desc, false,
            -2469mvcc);
            -2470  }
            -2471
            -2472  // Prepare flush (take a 
            snapshot)
            -2473  for (StoreFlushContext flush : 
            storeFlushCtxs.values()) {
            -2474flush.prepare();
            -2475  }
            -2476} catch (IOException ex) {
            -2477  doAbortFlushToWAL(wal, 
            flushOpSeqId, committedFiles);
            -2478  throw ex;
            -2479} finally {
            -2480  
            this.updatesLock.writeLock().unlock();
            -2481}
            -2482String s = "Finished memstore 
            snapshotting " + this + ", syncing WAL and waiting on mvcc, " +
            -2483"flushsize=" + 
            totalSizeOfFlushableStores;
            -2484status.setStatus(s);
            -2485doSyncOfUnflushedWALChanges(wal, 
            getRegionInfo());
            -2486return new 
            PrepareFlushResult(storeFlushCtxs, committedFiles, storeFlushableSize, 
            startTime,
            -2487flushOpSeqId, flushedSeqId, 
            totalSizeOfFlushableStores);
            -2488  }
            -2489
            -2490  /**
            -2491   * Utility method broken out of 
            internalPrepareFlushCache so that method is smaller.
            -2492   */
            -2493  private void 
            logFatLineOnFlush(CollectionHStore storesToFlush, long sequenceId) {
            -2494if (!LOG.isInfoEnabled()) {
            -2495  return;
            -2496}
            -2497// Log a fat line detailing what is 
            being flushed.
            -2498StringBuilder perCfExtras = null;
            -2499if (!isAllFamilies(storesToFlush)) 
            {
            -2500  perCfExtras = new 
            StringBuilder();
            -2501  for (HStore store: storesToFlush) 
            {
            -2502perCfExtras.append("; 
            ").append(store.getColumnFamilyName());
            -2503perCfExtras.append("=")
            -2504
            .append(StringUtils.byteDesc(store.getFlushableSize().getDataSize()));
            -2505  }
            -2506}
            -2507LOG.info("Flushing " + + 
            storesToFlush.size() + "/" + stores.size() +
            -2508" column families, memstore=" + 
            StringUtils.byteDesc(this.memstoreDataSize.get()) +
            -2509((perCfExtras != null  
            perCfExtras.length()  0)? perCfExtras.toString(): "") +
            -2510((wal != null) ? "" : "; WAL is 
            null, using passed sequenceid=" + sequenceId));
            -2511  }
            -2512
            -2513  private void doAbortFlushToWAL(final 
            WAL wal, final long flushOpSeqId,
            -2514  final Mapbyte[], 
            ListPath committedFiles) {
            -2515if (wal == null) return;
            -2516try {
            -2517  FlushDescriptor desc = 
            ProtobufUtil.toFlushDescriptor(FlushAction.ABORT_FLUSH,
            -2518  getRegionInfo(), flushOpSeqId, 
            committedFiles);
            -2519  WALUtil.writeFlushMarker(wal, 
            this.getReplicationScope(), getRegionInfo(), desc, false,
            -2520  mvcc);
            -2521} catch (Throwable t) {
            -2522  LOG.warn("Received unexpected 
            exception trying to write ABORT_FLUSH marker to WAL:" +
            -2523  
            StringUtils.stringifyException(t));
            -2524  // ignore this since we will be 
            aborting the RS with DSE.
            -2525}
            -2526// we have called 
            wal.startCacheFlush(), now we have to abort it
            -2527
            wal.abortCacheFlush(this.getRegionInfo().getEncodedNameAsBytes());
            -2528  }
            -2529
            -2530  /**
            -2531   * Sync unflushed WAL changes. See 
            HBASE-8208 for details
            -2532   */
            -2533  private static void 
            doSyncOfUnflushedWALChanges(final WAL wal, final RegionInfo hri)
            -2534  throws IOException {
            -2535if (wal == null) {
            -2536  return;
            -2537}
            -2538try {
            -2539  wal.sync(); // ensure that flush 
            marker is sync'ed
            -2540} catch (IOException ioe) {
            -2541  
            wal.abortCacheFlush(hri.getEncodedNameAsBytes());
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.html 
            b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.html
            index 560e87a..f61a992 100644
            --- a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.html
            +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.html
            @@ -18,7 +18,7 @@
             catch(err) {
             }
             //-->
            -var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":9,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":9,"i22":10,"i23":10,"i24":9};
            +var methods = 
            {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":9,"i16":10,"i17":10,"i18":10,"i19":10,"i20":9,"i21":10,"i22":10,"i23":9};
             var tabs = {65535:["t0","All Methods"],1:["t1","Static 
            Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
             var altColor = "altColor";
             var rowColor = "rowColor";
            @@ -400,39 +400,34 @@ extends 
             private SyncFuture
            -publishSyncOnRingBuffer(longsequence)
            +publishSyncOnRingBuffer()
             
             
             private SyncFuture
            -publishSyncOnRingBuffer(longsequence,
            -   org.apache.htrace.core.Spanspan)
            +publishSyncOnRingBuffer(longsequence)
             
             
            -private SyncFuture
            -publishSyncOnRingBuffer(org.apache.htrace.core.Spanspan)
            -
            -
             private void
             publishSyncThenBlockOnCompletion(org.apache.htrace.core.TraceScopescope)
             
            -
            +
             private static void
             split(org.apache.hadoop.conf.Configurationconf,
              org.apache.hadoop.fs.Pathp)
             
            -
            +
             void
             sync()
             Sync what we have in the WAL.
             
             
            -
            +
             void
             sync(longtxid)
             Sync the WAL if the txId was not already sync'd.
             
             
            -
            +
             private static void
             usage()
             
            @@ -442,7 +437,7 @@ extends AbstractFSWAL
            -abortCacheFlush,
             append,
             atHeadOfRingBufferEventHandlerAppend,
             blockOnSync,
             close,
             completeCacheFlush,
             c
             omputeFilename, findRegionsToForceFlush,
             getCoprocessorHost,
             getCurrentFileName,
             getEarliestMemStoreSeqNum,
             getEarliestMemStoreSeqNum,
             getFilenum,
             
             getFileNumFromFileName, getFiles,
             getLogFileSize,
             getLogFileSizeIfBeingWritten,
             getNumLogFiles,
             getNumRolledLogFiles,
             getOldPath,
             getPreallocatedEventCount,
             getSyncFuture,
             getUnflushedEntriesCount,
             getWALArchivePath,
             isUnflushedEntries,
             postSync,
             registerWALActionsListener,
             replaceWriter,
             requestLogRoll,
             requestLogRoll,
             rollWriter,
             rollWriter,
             shutdown,
             stampSequenc
             eIdAndPublishToRingBuffer, startCacheFlush,
             startCacheFlush,
             toString,
             unregisterWALActionsListener,
             updateStore
            +abortCacheFlush,
             append,
             atHeadOfRingBufferEventHandlerAppend,
             blockOnSync,
             close,
             completeCacheFlush,
             c
             omputeFilename, findRegionsToForceFlush,
             getCoprocessorHost,
             getCurrentFileName,
             getEarliestMemStoreSeqNum,
             getEarliestMemStoreSeqNum,
             getFilenum,
             
             getFileNumFromFileName, getFiles,
             getLogFileSize,
             getLogFileSizeIfBeingWritten,
             getNumLogFiles,
             getNumRolledLogFiles,
             getOldPath,
             getPreallocatedEventCount,
             getSyncFuture,
             getUnflushedEntriesCount,
             getWALArchivePath,
             isUnflushedEntries,
             postSync,
             registerWALActionsListener,
             replaceWriter, requestLogRoll,
             requestLogRoll,
             rollWriter,
             rollWriter,
             shutdown,
             stampSequenceIdAndPublishToRingBuffer, href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/AbstractFSWAL.html#startCacheFlush-byte:A-java.util.Map-">startCacheFlush,
             > href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/AbstractFSWAL.html#startCacheFlush-byte:A-java.util.Set-">startCacheFlush,
             > href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/AbstractFSWAL.html#toString--">toString,
             > href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/AbstractFSWAL.html#unregisterWALActionsListener-org.apache.hadoop.hbase.regionserver.wal.WALActionsListener-">unregisterWALActionsListener,
             > href="../../../../../../org/apache/hadoop/hbase/regionserver/wal/AbstractFSWAL.html#updateStore-byte:A-byte:A-java.lang.Long-boolean-">updateStore
             
             
             
            @@ -581,7 +576,7 @@ extends 
             
             FIXED_OVERHEAD
            -public static finallong FIXED_OVERHEAD
            +public static finallong FIXED_OVERHEAD
             
             
             
            @@ -856,41 +851,31 @@ extends 
            -
            -
            -
            -
            -publishSyncOnRingBuffer
            -privateSyncFuturepublishSyncOnRingBuffer(longsequence)
            -
            -
             
             
             
             
             
             getSequenceOnRingBuffer
            -privatelonggetSequenceOnRingBuffer()
            +privatelonggetSequenceOnRingBuffer()
             
             
            -
            +
             
             
             
             
             publishSyncOnRingBuffer
            -privateSyncFuturepublishSyncOnRingBuffer(org.apache.htrace.core.Spanspan)
            +privateSyncFuturepublishSyncOnRingBuffer()
             
             
            -
            +
             
             
             
             
             publishSyncOnRingBuffer
            -privateSyncFuturepublishSyncOnRingBuffer(longsequence,
            -   
            org.apache.htrace.core.Spanspan)
            +privateSyncFuturepublishSyncOnRingBuffer(longsequence)
             
             
             
            @@ -899,7 +884,7 @@ extends 
             
             

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
            --
            diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html 
            b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
            index 97e977c..c8967e5 100644
            --- a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
            +++ b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
            @@ -543,23 +543,23 @@
             java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
             title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
             title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
             title="class or interface in java.io">Serializable)
             
             org.apache.hadoop.hbase.client.ScannerCallable.MoreResults
            -org.apache.hadoop.hbase.client.SnapshotType
            -org.apache.hadoop.hbase.client.AsyncRequestFutureImpl.Retry
            -org.apache.hadoop.hbase.client.RequestController.ReturnCode
            -org.apache.hadoop.hbase.client.MobCompactPartitionPolicy
            -org.apache.hadoop.hbase.client.RegionLocateType
            +org.apache.hadoop.hbase.client.MasterSwitchType
             org.apache.hadoop.hbase.client.CompactType
             org.apache.hadoop.hbase.client.Scan.ReadType
            -org.apache.hadoop.hbase.client.IsolationLevel
            -org.apache.hadoop.hbase.client.CompactionState
            +org.apache.hadoop.hbase.client.RegionLocateType
            +org.apache.hadoop.hbase.client.SnapshotType
            +org.apache.hadoop.hbase.client.MobCompactPartitionPolicy
            +org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState
            +org.apache.hadoop.hbase.client.Durability
             org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState
            +org.apache.hadoop.hbase.client.CompactionState
            +org.apache.hadoop.hbase.client.RequestController.ReturnCode
            +org.apache.hadoop.hbase.client.IsolationLevel
             org.apache.hadoop.hbase.client.Consistency
            -org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState
             org.apache.hadoop.hbase.client.AbstractResponse.ResponseType
            -org.apache.hadoop.hbase.client.MasterSwitchType
             org.apache.hadoop.hbase.client.TableState.State
             org.apache.hadoop.hbase.client.AsyncProcessTask.SubmittedRows
            -org.apache.hadoop.hbase.client.Durability
            +org.apache.hadoop.hbase.client.AsyncRequestFutureImpl.Retry
             
             
             
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.GetDataAsyncCallback.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.GetDataAsyncCallback.html
             
            b/devapidocs/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.GetDataAsyncCallback.html
            index f6dbdb5..ebf2337 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.GetDataAsyncCallback.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.GetDataAsyncCallback.html
            @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -class ZkSplitLogWorkerCoordination.GetDataAsyncCallback
            +class ZkSplitLogWorkerCoordination.GetDataAsyncCallback
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
             implements org.apache.zookeeper.AsyncCallback.DataCallback
             Asynchronous handler for zk get-data-set-watch on node 
            results.
            @@ -226,7 +226,7 @@ implements 
            org.apache.zookeeper.AsyncCallback.DataCallback
             
             
             LOG
            -private finalorg.apache.commons.logging.Log LOG
            +private finalorg.apache.commons.logging.Log LOG
             
             
             
            @@ -243,7 +243,7 @@ implements 
            org.apache.zookeeper.AsyncCallback.DataCallback
             
             
             GetDataAsyncCallback
            -GetDataAsyncCallback()
            +GetDataAsyncCallback()
             
             
             
            @@ -260,7 +260,7 @@ implements 
            org.apache.zookeeper.AsyncCallback.DataCallback
             
             
             processResult
            -publicvoidprocessResult(intrc,
            +publicvoidprocessResult(intrc,
               http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
             title="class or interface in java.lang">Stringpath,
               http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Objectctx,
               byte[]data,
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.ZkSplitTaskDetails.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/coordination/ZkSplitLogWorkerCoordination.ZkSplitTaskDetails.html
             
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9722a17/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            --
            diff --git 
            a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
             
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            index 51844c3..bbd7a35 100644
            --- 
            a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            +++ 
            b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
            @@ -55,2130 +55,2170 @@
             047import 
            org.apache.hadoop.hbase.client.TableState;
             048import 
            org.apache.hadoop.hbase.client.VersionInfoUtil;
             049import 
            org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil;
            -050import 
            org.apache.hadoop.hbase.errorhandling.ForeignException;
            -051import 
            org.apache.hadoop.hbase.exceptions.UnknownProtocolException;
            -052import 
            org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
            -053import 
            org.apache.hadoop.hbase.ipc.PriorityFunction;
            -054import 
            org.apache.hadoop.hbase.ipc.QosPriority;
            -055import 
            org.apache.hadoop.hbase.ipc.RpcServer.BlockingServiceAndInterface;
            -056import 
            org.apache.hadoop.hbase.ipc.ServerRpcController;
            -057import 
            org.apache.hadoop.hbase.master.assignment.RegionStates;
            -058import 
            org.apache.hadoop.hbase.master.locking.LockProcedure;
            -059import 
            org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
            -060import 
            org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil.NonceProcedureRunnable;
            -061import 
            org.apache.hadoop.hbase.mob.MobUtils;
            -062import 
            org.apache.hadoop.hbase.procedure.MasterProcedureManager;
            -063import 
            org.apache.hadoop.hbase.procedure2.LockType;
            -064import 
            org.apache.hadoop.hbase.procedure2.LockedResource;
            -065import 
            org.apache.hadoop.hbase.procedure2.Procedure;
            -066import 
            org.apache.hadoop.hbase.procedure2.ProcedureUtil;
            -067import 
            org.apache.hadoop.hbase.quotas.MasterQuotaManager;
            -068import 
            org.apache.hadoop.hbase.quotas.QuotaObserverChore;
            -069import 
            org.apache.hadoop.hbase.quotas.QuotaUtil;
            -070import 
            org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshot;
            -071import 
            org.apache.hadoop.hbase.regionserver.RSRpcServices;
            -072import 
            org.apache.hadoop.hbase.replication.ReplicationException;
            -073import 
            org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
            -074import 
            org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
            -075import 
            org.apache.hadoop.hbase.security.User;
            -076import 
            org.apache.hadoop.hbase.security.access.AccessController;
            -077import 
            org.apache.hadoop.hbase.security.visibility.VisibilityController;
            -078import 
            org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
            -079import 
            org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils;
            -080import 
            org.apache.hadoop.hbase.util.Bytes;
            -081import 
            org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
            -082import 
            org.apache.hadoop.hbase.util.ForeignExceptionUtil;
            -083import 
            org.apache.hadoop.hbase.util.Pair;
            -084import 
            org.apache.yetus.audience.InterfaceAudience;
            -085import 
            org.apache.zookeeper.KeeperException;
            -086import 
            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController;
            -087import 
            org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
            -088import 
            org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
            -089import 
            org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
            -090import 
            org.apache.hadoop.hbase.shaded.protobuf.ResponseConverter;
            -091import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
            -092import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse;
            -093import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
            -094import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
            -095import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
            -096import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos;
            -097import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds;
            -098import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
            -099import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair;
            -100import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
            -101import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
            -102import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockHeartbeatRequest;
            -103import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockHeartbeatResponse;
            -104import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockRequest;
            -105import 
            org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockResponse;
            

            [26/51] [partial] hbase-site git commit: Published site at .

            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b1eb7453/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OnheapDecodedCell.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OnheapDecodedCell.html
             
            b/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OnheapDecodedCell.html
            index f64a4fa..11b96bc 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OnheapDecodedCell.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.OnheapDecodedCell.html
            @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab";
             
             
             All Implemented Interfaces:
            -http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true;
             title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, SettableSequenceId, SettableTimestamp
            +http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true;
             title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell, SettableSequenceId, SettableTimestamp
             
             
             Enclosing class:
            @@ -214,6 +214,13 @@ implements Fields inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell
             CELL_NOT_BASED_ON_CHUNK
             
            +
            +
            +
            +
            +Fields inherited from interfaceorg.apache.hadoop.hbase.RawCell
            +MAX_TAGS_LENGTH
            +
             
             
             
            @@ -414,6 +421,13 @@ implements Methods inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell
             getChunkId
             
            +
            +
            +
            +
            +Methods inherited from interfaceorg.apache.hadoop.hbase.RawCell
            +checkForTagsLength,
             cloneTags,
             getTag,
             getTags
            +
             
             
             
            @@ -895,9 +909,10 @@ implements Description copied from 
            interface:Cell
             HBase internally uses 2 bytes to store tags length in Cell.
              As the tags length is always a non-negative number, to make good use of the 
            sign bit,
            - the max of tags length is defined as TagUtil.MAX_TAGS_LENGTH,
             which is 2 * Short.MAX_VALUE + 1 = 65535.
            + the max of tags length is defined 2 * Short.MAX_VALUE + 1 = 65535.
              As a result, the return type is int, because a short is not capable of 
            handling that.
            - Please note that even if the return type is int, the max tags length is far 
            less than Integer.MAX_VALUE.
            + Please note that even if the return type is int, the max tags length is far
            + less than Integer.MAX_VALUE.
             
             Specified by:
             getTagsLengthin
             interfaceCell
            
            http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b1eb7453/devapidocs/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.KeyValueStatsCollector.html
            --
            diff --git 
            a/devapidocs/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.KeyValueStatsCollector.html
             
            b/devapidocs/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.KeyValueStatsCollector.html
            index 95f33e4..cf9843c 100644
            --- 
            a/devapidocs/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.KeyValueStatsCollector.html
            +++ 
            b/devapidocs/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.KeyValueStatsCollector.html
            @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
             
             
             
            -private static class HFilePrettyPrinter.KeyValueStatsCollector
            +private static class HFilePrettyPrinter.KeyValueStatsCollector
             extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
             title="class or interface in java.lang">Object
             
             
            @@ -263,7 +263,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             metricsRegistry
            -private finalcom.codahale.metrics.MetricRegistry metricsRegistry
            +private finalcom.codahale.metrics.MetricRegistry metricsRegistry
             
             
             
            @@ -272,7 +272,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             metricsOutput
            -private finalhttp://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true;
             title="class or interface in java.io">ByteArrayOutputStream metricsOutput
            +private finalhttp://docs.oracle.com/javase/8/docs/api/java/io/ByteArrayOutputStream.html?is-external=true;
             title="class or interface in java.io">ByteArrayOutputStream metricsOutput
             
             
             
            @@ -281,7 +281,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             simpleReporter
            -private finalHFilePrettyPrinter.SimpleReporter simpleReporter
            +private finalHFilePrettyPrinter.SimpleReporter simpleReporter
             
             
             
            @@ -290,7 +290,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             keyLen
            -com.codahale.metrics.Histogram keyLen
            +com.codahale.metrics.Histogram keyLen
             
             
             
            @@ -299,7 +299,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             valLen
            -com.codahale.metrics.Histogram valLen
            +com.codahale.metrics.Histogram valLen
             
             
             
            @@ -308,7 +308,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
             
             
             rowSizeBytes
            -com.codahale.metrics.Histogram rowSizeBytes
            +com.codahale.metrics.Histogram rowSizeBytes
             
             
             
            @@ -317,7 +317,7 @@ 

              1   2   3   >