[23/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.

2019-01-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
index 68302bf..a5a8905 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
@@ -2197,1768 +2197,1775 @@
 2189  
warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null);
 2190}
 2191
-2192for (ColumnFamilyDescriptor hcd : 
htd.getColumnFamilies()) {
-2193  if (hcd.getTimeToLive() = 0) 
{
-2194String message = "TTL for column 
family " + hcd.getNameAsString() + " must be positive.";
-2195
warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null);
-2196  }
-2197
-2198  // check blockSize
-2199  if (hcd.getBlocksize()  1024 
|| hcd.getBlocksize()  16 * 1024 * 1024) {
-2200String message = "Block size for 
column family " + hcd.getNameAsString()
-2201+ "  must be between 1K and 
16MB.";
+2192// check that we have minimum 1 
region replicas
+2193int regionReplicas = 
htd.getRegionReplication();
+2194if (regionReplicas  1) {
+2195  String message = "Table region 
replication should be at least one.";
+2196  
warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null);
+2197}
+2198
+2199for (ColumnFamilyDescriptor hcd : 
htd.getColumnFamilies()) {
+2200  if (hcd.getTimeToLive() = 0) 
{
+2201String message = "TTL for column 
family " + hcd.getNameAsString() + " must be positive.";
 2202
warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null);
 2203  }
 2204
-2205  // check versions
-2206  if (hcd.getMinVersions()  0) 
{
-2207String message = "Min versions 
for column family " + hcd.getNameAsString()
-2208  + "  must be positive.";
+2205  // check blockSize
+2206  if (hcd.getBlocksize()  1024 
|| hcd.getBlocksize()  16 * 1024 * 1024) {
+2207String message = "Block size for 
column family " + hcd.getNameAsString()
+2208+ "  must be between 1K and 
16MB.";
 2209
warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null);
 2210  }
-2211  // max versions already being 
checked
-2212
-2213  // HBASE-13776 Setting illegal 
versions for ColumnFamilyDescriptor
-2214  //  does not throw 
IllegalArgumentException
-2215  // check minVersions = 
maxVerions
-2216  if (hcd.getMinVersions()  
hcd.getMaxVersions()) {
-2217String message = "Min versions 
for column family " + hcd.getNameAsString()
-2218+ " must be less than the 
Max versions.";
-2219
warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null);
-2220  }
-2221
-  // check replication scope
-2223  checkReplicationScope(hcd);
-2224  // check bloom filter type
-2225  checkBloomFilterType(hcd);
-2226
-2227  // check data replication factor, 
it can be 0(default value) when user has not explicitly
-2228  // set the value, in this case we 
use default replication factor set in the file system.
-2229  if (hcd.getDFSReplication()  
0) {
-2230String message = "HFile 
Replication for column family " + hcd.getNameAsString()
-2231+ "  must be greater than 
zero.";
-2232
warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null);
-2233  }
-2234
-2235  // TODO: should we check 
coprocessors and encryption ?
-2236}
-2237  }
-2238
-2239  private void 
checkReplicationScope(ColumnFamilyDescriptor hcd) throws IOException{
-2240// check replication scope
-2241WALProtos.ScopeType scop = 
WALProtos.ScopeType.valueOf(hcd.getScope());
-2242if (scop == null) {
-2243  String message = "Replication 
scope for column family "
-2244  + hcd.getNameAsString() + " is 
" + hcd.getScope() + " which is invalid.";
+2211
+2212  // check versions
+2213  if (hcd.getMinVersions()  0) 
{
+2214String message = "Min versions 
for column family " + hcd.getNameAsString()
+2215  + "  must be positive.";
+2216
warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null);
+2217  }
+2218  // max versions already being 
checked
+2219
+2220  // HBASE-13776 Setting illegal 
versions for ColumnFamilyDescriptor
+2221  //  does not throw 
IllegalArgumentException
+  // check minVersions = 
maxVerions
+2223  if (hcd.getMinVersions()  
hcd.getMaxVersions()) {
+2224String message = "Min versions 
for column family " + hcd.getNameAsString()
+2225+ " must be less than the 
Max versions.";
+2226
warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, 

[23/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.

2019-01-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
 
b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
index 60a5382..5e9a1d4 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
@@ -247,7 +247,7 @@ extends 
 
 Methods inherited from classorg.apache.hadoop.hbase.master.HMaster
-abort,
 abortProcedure,
 addColumn,
 addReplicationPeer,
 balance,
 balance,
 balanceSwitch,
 canCreateBaseZNode, canUpdateTableDescriptor,
 checkIfShouldMoveSystemRegionAsync,
 checkInitialized,
 checkServiceStarted,
 checkTableModifiable,
 configureInfoServer,
 constructMaster,
 createAssignmentManager,
 createMetaBootstrap,
 createNamespace,
 createRpcServices,
 createServerManager,
 createSystemTable,
 createTable, decommissionRegionServers,
 decorateMasterConfiguration,
 deleteColumn,
 deleteNamespace,
 deleteTable,
 disableReplicationPeer,
 disableTable, enableReplicationPeer,
 enableTable,
 getAssignmentManager,
 getAverageLoad,
 getCatalogJanitor,
 getClientIdAuditPrefix,
 getClusterMetrics,
 getClusterMetrics, getClusterMetricsWithoutCoprocessor,
 getClusterMetricsWithoutCoprocessor,
 getClusterSchema,
 getDumpServlet,
 getFavoredNodesManager,
 getHFileCleaner,
 getInitializedEvent,
 getLastMajorCompactionTimestamp,
 getLastMajorCompactionTimestampForRegion,
 getLoadBalancer,
 getLoadBalancerClassName,
 getLoadedCoprocessors,
 getLockManager,
 getLocks,
 getLogCleaner,
 getMasterActiveTime, getMasterCoprocessorHost,
 getMasterCoprocessors,
 getMasterFileSystem,
 getMasterFinishedInitializationTime,
 getMasterMetrics,
 getMasterProcedureExecutor,
 getMasterProcedureManagerHost,
 getMasterQuotaManager, getMasterRpcServices,
 getMasterStartTime,
 getMasterWalManager,
 getMergePlanCount,
 getMetaTableObserver,
 getMobCompactionState,
 getNamespace,
 getNamespaces,
 
 getNumWALFiles,
 getProcedures,
 getProcessName,
 getQuotaObserverChore,
 getRegionNormalizer,
 getRegionNormalizerTracker,
 getRegionServerFatalLogBuffer,
 getRegionServerInfoPort,
 getRegionServerVersion,
 getRemoteInetAddress,
 getReplicationLoad,
 getReplicationPeerConfig,
 getReplicationPeerManager,
 getServerManager,
 getServerName,
 getSnapshotM
 anager, getSnapshotQuotaObserverChore,
 getSpaceQuotaSnapshotNotifier,
 getSplitOrMergeTracker,
 getSplitPlanCount,
 getSyncReplicationReplayWALManager,
 getTableStateManager,
 getUseThisHostnameInstead,
 getWalProcedureStore, getZooKeeper,
 initClusterSchemaService,
 initializeZKBasedSystemTrackers,
 isActiveMaster,
 isBalancerOn,
 isCatalogJanitorEnabled,
 isCleanerChoreEnabled,
 isInitialized,
 isInMaintenanceMode,
 isNormalizerOn,
 isSplitOrMergeEnabled,
 listDecommissionedRegionServers,
 listReplicationPeers,
 listTableDescriptors,
 listTableDescriptorsByNamespace,
 listTableNames,
 listTableNamesByNamespace,
 login,
 main,
 mergeRegions,
 modifyColumn,
 modifyNamespace, modifyTable,
 move,
 normalizeRegions,
 recommissionRegionServer,
 registerService,
 remoteProcedureCompleted,
 remoteProcedureFailed,
 removeReplicationPeer,
 reportMobCompactionEnd,
 reportMobCompactionStart,
 requestMobCompaction,
 restoreSnapshot,
 setCatalogJanitorEnabled, setInitialized,
 shutdown,
 splitRegion,
 stop,
 stopMaster,
 stopServiceThreads,
 transitReplicationPeerSyncReplicationState,
 truncateTable,
 updateConfigurationForQuotasObserver,
 updateReplicationPeerConfig,
 waitForMasterActive,
 waitForMetaOnline
+abort,
 abortProcedure,
 addColumn,
 addReplicationPeer,
 balance,
 balance,
 balanceSwitch,
 canCreateBaseZNode, canUpdateTableDescriptor,
 checkIfShouldMoveSystemRegionAsync,
 checkInitialized,
 checkServiceStarted,
 checkTableModifiable,
 configureInfoServer,
 constructMaster,
 createAssignmentManager,
 createMetaBootstrap,
 createNamespace,
 createRpcServices,
 createServerManager,
 createSystemTable,
 createTable, decommissionRegionServers,
 decorateMasterConfiguration,
 deleteColumn,
 deleteNamespace,
 deleteTable,
 disableReplicationPeer,
 disableTable, enableReplicationPeer,
 enableTable,
 getAssignmentManager,
 getAverageLoad,
 getCatalogJanitor,
 getClientIdAuditPrefix,
 getClusterMetrics,
 getClusterMetrics, getClusterMetricsWithoutCoprocessor,
 getClusterMetricsWithoutCoprocessor,
 getClusterSchema,
 getDumpServlet,
 getFavoredNodesManager,
 getHFileCleaner,
 getInitializedEvent,
 getLastMajorCompactionTimestamp,
 getLastMajorCompactionTimestampForRegion,
 getLoadBalancer,
 getLoadBalancerClassName,
 getLoadedCoprocessors,
 getLockManager,
 getLocks,
 getLogCleaner,
 getMasterActiveTime, getMasterCoprocessorHost,
 getMasterCoprocessors,
 getMasterFileSystem,
 getMasterFinishedInitializationTime,
 

[23/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.

2019-01-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html
index 9aa9b59..ac7e0ea 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html
@@ -28,520 +28,565 @@
 020import static 
org.apache.hadoop.hbase.HConstants.NINES;
 021import static 
org.apache.hadoop.hbase.HConstants.ZEROES;
 022import static 
org.apache.hadoop.hbase.TableName.META_TABLE_NAME;
-023import static 
org.apache.hadoop.hbase.client.ConnectionUtils.createClosestRowAfter;
-024import static 
org.apache.hadoop.hbase.client.ConnectionUtils.isEmptyStopRow;
-025import static 
org.apache.hadoop.hbase.client.RegionInfo.createRegionName;
-026import static 
org.apache.hadoop.hbase.util.Bytes.BYTES_COMPARATOR;
-027import static 
org.apache.hadoop.hbase.util.CollectionUtils.computeIfAbsent;
-028
-029import java.io.IOException;
-030import java.util.Arrays;
-031import java.util.HashSet;
-032import java.util.Iterator;
-033import java.util.LinkedHashMap;
-034import java.util.Map;
-035import java.util.Optional;
-036import java.util.Set;
-037import 
java.util.concurrent.CompletableFuture;
-038import 
java.util.concurrent.ConcurrentHashMap;
-039import 
java.util.concurrent.ConcurrentMap;
-040import 
java.util.concurrent.ConcurrentNavigableMap;
-041import 
java.util.concurrent.ConcurrentSkipListMap;
-042import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-043import 
org.apache.hadoop.hbase.HConstants;
-044import 
org.apache.hadoop.hbase.HRegionLocation;
-045import 
org.apache.hadoop.hbase.MetaTableAccessor;
-046import 
org.apache.hadoop.hbase.RegionLocations;
-047import 
org.apache.hadoop.hbase.TableName;
-048import 
org.apache.hadoop.hbase.TableNotFoundException;
-049import 
org.apache.hadoop.hbase.client.Scan.ReadType;
-050import 
org.apache.hadoop.hbase.util.Bytes;
-051import 
org.apache.yetus.audience.InterfaceAudience;
-052import org.slf4j.Logger;
-053import org.slf4j.LoggerFactory;
-054
-055import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-056
-057/**
-058 * The asynchronous locator for regions 
other than meta.
-059 */
-060@InterfaceAudience.Private
-061class AsyncNonMetaRegionLocator {
-062
-063  private static final Logger LOG = 
LoggerFactory.getLogger(AsyncNonMetaRegionLocator.class);
+023import static 
org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.canUpdateOnError;
+024import static 
org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.createRegionLocations;
+025import static 
org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.isGood;
+026import static 
org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.mergeRegionLocations;
+027import static 
org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.removeRegionLocation;
+028import static 
org.apache.hadoop.hbase.client.ConnectionUtils.createClosestRowAfter;
+029import static 
org.apache.hadoop.hbase.client.ConnectionUtils.isEmptyStopRow;
+030import static 
org.apache.hadoop.hbase.client.RegionInfo.createRegionName;
+031import static 
org.apache.hadoop.hbase.util.Bytes.BYTES_COMPARATOR;
+032import static 
org.apache.hadoop.hbase.util.CollectionUtils.computeIfAbsent;
+033
+034import java.io.IOException;
+035import java.util.Arrays;
+036import java.util.HashSet;
+037import java.util.Iterator;
+038import java.util.LinkedHashMap;
+039import java.util.Map;
+040import java.util.Optional;
+041import java.util.Set;
+042import 
java.util.concurrent.CompletableFuture;
+043import 
java.util.concurrent.ConcurrentHashMap;
+044import 
java.util.concurrent.ConcurrentMap;
+045import 
java.util.concurrent.ConcurrentNavigableMap;
+046import 
java.util.concurrent.ConcurrentSkipListMap;
+047import 
org.apache.commons.lang3.ObjectUtils;
+048import 
org.apache.hadoop.hbase.DoNotRetryIOException;
+049import 
org.apache.hadoop.hbase.HBaseIOException;
+050import 
org.apache.hadoop.hbase.HConstants;
+051import 
org.apache.hadoop.hbase.HRegionLocation;
+052import 
org.apache.hadoop.hbase.MetaTableAccessor;
+053import 
org.apache.hadoop.hbase.RegionLocations;
+054import 
org.apache.hadoop.hbase.TableName;
+055import 
org.apache.hadoop.hbase.TableNotFoundException;
+056import 
org.apache.hadoop.hbase.client.Scan.ReadType;
+057import 
org.apache.hadoop.hbase.util.Bytes;
+058import 
org.apache.yetus.audience.InterfaceAudience;
+059import org.slf4j.Logger;
+060import org.slf4j.LoggerFactory;
+061
+062import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+063import 
org.apache.hbase.thirdparty.com.google.common.base.Objects;
 064
-065  @VisibleForTesting
-066  static final 

[23/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.

2019-01-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/src-html/org/apache/hadoop/hbase/thrift/Constants.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/Constants.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/Constants.html
new file mode 100644
index 000..d28d77e
--- /dev/null
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/Constants.html
@@ -0,0 +1,223 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/*
+002 * Licensed to the Apache Software 
Foundation (ASF) under one
+003 * or more contributor license 
agreements.  See the NOTICE file
+004 * distributed with this work for 
additional information
+005 * regarding copyright ownership.  The 
ASF licenses this file
+006 * to you under the Apache License, 
Version 2.0 (the
+007 * "License"); you may not use this file 
except in compliance
+008 * with the License.  You may obtain a 
copy of the License at
+009 *
+010 * 
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or 
agreed to in writing, software
+013 * distributed under the License is 
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+015 * See the License for the specific 
language governing permissions and
+016 * limitations under the License.
+017 */
+018package org.apache.hadoop.hbase.thrift;
+019
+020import 
org.apache.yetus.audience.InterfaceAudience;
+021
+022/**
+023 * Thrift related constants
+024 */
+025@InterfaceAudience.Private
+026public final class Constants {
+027  private Constants(){}
+028
+029  public static final int 
DEFAULT_HTTP_MAX_HEADER_SIZE = 64 * 1024; // 64k
+030
+031  public static final String 
SERVER_TYPE_CONF_KEY =
+032  
"hbase.regionserver.thrift.server.type";
+033
+034  public static final String 
COMPACT_CONF_KEY = "hbase.regionserver.thrift.compact";
+035  public static final boolean 
COMPACT_CONF_DEFAULT = false;
+036
+037  public static final String 
FRAMED_CONF_KEY = "hbase.regionserver.thrift.framed";
+038  public static final boolean 
FRAMED_CONF_DEFAULT = false;
+039
+040  public static final String 
MAX_FRAME_SIZE_CONF_KEY =
+041  
"hbase.regionserver.thrift.framed.max_frame_size_in_mb";
+042  public static final int 
MAX_FRAME_SIZE_CONF_DEFAULT = 2;
+043
+044  public static final String 
COALESCE_INC_KEY = "hbase.regionserver.thrift.coalesceIncrement";
+045  public static final String 
USE_HTTP_CONF_KEY = "hbase.regionserver.thrift.http";
+046
+047  public static final String 
HTTP_MIN_THREADS_KEY = "hbase.thrift.http_threads.min";
+048  public static final int 
HTTP_MIN_THREADS_KEY_DEFAULT = 2;
+049
+050  public static final String 
HTTP_MAX_THREADS_KEY = "hbase.thrift.http_threads.max";
+051  public static final int 
HTTP_MAX_THREADS_KEY_DEFAULT = 100;
+052
+053  // ssl related configs
+054  public static final String 
THRIFT_SSL_ENABLED_KEY = "hbase.thrift.ssl.enabled";
+055  public static final String 
THRIFT_SSL_KEYSTORE_STORE_KEY = "hbase.thrift.ssl.keystore.store";
+056  public static final String 
THRIFT_SSL_KEYSTORE_PASSWORD_KEY =
+057  
"hbase.thrift.ssl.keystore.password";
+058  public static final String 
THRIFT_SSL_KEYSTORE_KEYPASSWORD_KEY
+059  = 
"hbase.thrift.ssl.keystore.keypassword";
+060  public static final String 
THRIFT_SSL_EXCLUDE_CIPHER_SUITES_KEY =
+061  
"hbase.thrift.ssl.exclude.cipher.suites";
+062  public static final String 
THRIFT_SSL_INCLUDE_CIPHER_SUITES_KEY =
+063  
"hbase.thrift.ssl.include.cipher.suites";
+064  public static final String 
THRIFT_SSL_EXCLUDE_PROTOCOLS_KEY =
+065  
"hbase.thrift.ssl.exclude.protocols";
+066  public static final String 
THRIFT_SSL_INCLUDE_PROTOCOLS_KEY =
+067  
"hbase.thrift.ssl.include.protocols";
+068
+069
+070  public static final String 
THRIFT_SUPPORT_PROXYUSER_KEY = "hbase.thrift.support.proxyuser";
+071
+072  //kerberos related configs
+073  public static final String 
THRIFT_DNS_INTERFACE_KEY = "hbase.thrift.dns.interface";
+074  public static final String 
THRIFT_DNS_NAMESERVER_KEY = "hbase.thrift.dns.nameserver";
+075  public static final String 
THRIFT_KERBEROS_PRINCIPAL_KEY = "hbase.thrift.kerberos.principal";
+076  public static final String 
THRIFT_KEYTAB_FILE_KEY = "hbase.thrift.keytab.file";
+077  public static final String 
THRIFT_SPNEGO_PRINCIPAL_KEY = "hbase.thrift.spnego.principal";
+078  public static final String 
THRIFT_SPNEGO_KEYTAB_FILE_KEY = "hbase.thrift.spnego.keytab.file";
+079
+080  /**
+081   * Amount of time in milliseconds 
before a server thread will timeout
+082   * waiting for client to send data on a 
connected socket. Currently,
+083   * applies only to 
TBoundedThreadPoolServer
+084   */
+085  public static final String 
THRIFT_SERVER_SOCKET_READ_TIMEOUT_KEY =
+086  
"hbase.thrift.server.socket.read.timeout";
+087  public static final int 

[23/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.

2019-01-01 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html 
b/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html
index 14dcf5b..16c4959 100644
--- a/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html
+++ b/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html
@@ -275,6 +275,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html
index 4468eff..eb81cc7 100644
--- a/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html
@@ -204,6 +204,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/codec/package-use.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/codec/package-use.html 
b/devapidocs/org/apache/hadoop/hbase/codec/package-use.html
index 76343f5..41324f1 100644
--- a/devapidocs/org/apache/hadoop/hbase/codec/package-use.html
+++ b/devapidocs/org/apache/hadoop/hbase/codec/package-use.html
@@ -277,6 +277,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html 
b/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html
index 6ea799e..34baa82 100644
--- a/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html
+++ b/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html
@@ -420,6 +420,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html 
b/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html
index 72b04d4..d588c0a 100644
--- a/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html
+++ b/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html
@@ -235,6 +235,6 @@ public interface Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html 
b/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html
index 880a869..6896ab9 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html
@@ -265,6 +265,6 @@ extends Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/conf/class-use/ConfigurationManager.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/conf/class-use/ConfigurationManager.html 
b/devapidocs/org/apache/hadoop/hbase/conf/class-use/ConfigurationManager.html
index 143db46..960e9f4 100644
--- 

[23/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.

2018-12-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
index 4ade4c1..03e1383 100644
--- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
@@ -199,9 +199,9 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
+org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage
 org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType
 org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType
-org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html
 
b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html
index 8b8ba09..a21f2ad 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static final class ThriftHBaseServiceHandler.THBaseServiceMetricsProxy
+private static final class ThriftHBaseServiceHandler.THBaseServiceMetricsProxy
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/InvocationHandler.html?is-external=true;
 title="class or interface in java.lang.reflect">InvocationHandler
 
@@ -215,7 +215,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/
 
 
 handler
-private 
finalorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Iface handler
+private 
finalorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Iface handler
 
 
 
@@ -224,7 +224,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/
 
 
 metrics
-private finalThriftMetrics metrics
+private finalThriftMetrics metrics
 
 
 
@@ -241,7 +241,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/
 
 
 THBaseServiceMetricsProxy
-privateTHBaseServiceMetricsProxy(org.apache.hadoop.hbase.thrift2.generated.THBaseService.Ifacehandler,
+privateTHBaseServiceMetricsProxy(org.apache.hadoop.hbase.thrift2.generated.THBaseService.Ifacehandler,
   ThriftMetricsmetrics)
 
 
@@ -259,7 +259,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/
 
 
 invoke
-publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objectinvoke(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objectproxy,
+publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objectinvoke(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objectproxy,
  https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true;
 title="class or interface in java.lang.reflect">Methodm,
  https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object[]args)
   throws https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwable

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.TIOErrorWithCause.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.TIOErrorWithCause.html
 
b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.TIOErrorWithCause.html
index 24184b7..bf9ac4e 100644
--- 

[23/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.

2018-12-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
index 84c4071..e7b6097 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
@@ -105,7 +105,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-static interface RSRpcServices.LogDelegate
+static interface RSRpcServices.LogDelegate
 
 
 
@@ -151,7 +151,7 @@ var activeTableTab = "activeTableTab";
 
 
 logBatchWarning
-voidlogBatchWarning(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringfirstRegionName,
+voidlogBatchWarning(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringfirstRegionName,
  intsum,
  introwSizeWarnThreshold)
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
index 5d0e72a..9e736dc 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static final class RSRpcServices.RegionScannerCloseCallBack
+private static final class RSRpcServices.RegionScannerCloseCallBack
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements RpcCallback
 An Rpc callback for closing a RegionScanner.
@@ -209,7 +209,7 @@ implements 
 
 scanner
-private finalRegionScanner scanner
+private finalRegionScanner scanner
 
 
 
@@ -226,7 +226,7 @@ implements 
 
 RegionScannerCloseCallBack
-publicRegionScannerCloseCallBack(RegionScannerscanner)
+publicRegionScannerCloseCallBack(RegionScannerscanner)
 
 
 
@@ -243,7 +243,7 @@ implements 
 
 run
-publicvoidrun()
+publicvoidrun()
  throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Description copied from 
interface:RpcCallback
 Called at the end of an Rpc Call RpcCallContext

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
index 70b0bd0..69d484d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static final class RSRpcServices.RegionScannerHolder
+private static final class RSRpcServices.RegionScannerHolder
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 Holder class which holds the RegionScanner, nextCallSeq and 
RpcCallbacks together.
 
@@ -239,7 +239,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 nextCallSeq
-private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true;
 title="class or interface in java.util.concurrent.atomic">AtomicLong nextCallSeq
+private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true;
 title="class or interface in java.util.concurrent.atomic">AtomicLong nextCallSeq
 
 
 
@@ -248,7 +248,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 scannerName
-private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String scannerName
+private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String scannerName
 
 
 
@@ -257,7 +257,7 @@ extends 

[23/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.

2018-12-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html
index ea05301..26a93dd 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html
@@ -269,3590 +269,3574 @@
 261   */
 262  protected ClusterConnection 
clusterConnection;
 263
-264  /*
-265   * Long-living meta table locator, 
which is created when the server is started and stopped
-266   * when server shuts down. References 
to this locator shall be used to perform according
-267   * operations in EventHandlers. Primary 
reason for this decision is to make it mockable
-268   * for tests.
-269   */
-270  protected MetaTableLocator 
metaTableLocator;
-271
-272  /**
-273   * Go here to get table descriptors.
-274   */
-275  protected TableDescriptors 
tableDescriptors;
-276
-277  // Replication services. If no 
replication, this handler will be null.
-278  protected ReplicationSourceService 
replicationSourceHandler;
-279  protected ReplicationSinkService 
replicationSinkHandler;
-280
-281  // Compactions
-282  public CompactSplit 
compactSplitThread;
-283
-284  /**
-285   * Map of regions currently being 
served by this region server. Key is the
-286   * encoded region name.  All access 
should be synchronized.
-287   */
-288  protected final MapString, 
HRegion onlineRegions = new ConcurrentHashMap();
-289
-290  /**
-291   * Map of encoded region names to the 
DataNode locations they should be hosted on
-292   * We store the value as 
InetSocketAddress since this is used only in HDFS
-293   * API (create() that takes favored 
nodes as hints for placing file blocks).
-294   * We could have used ServerName here 
as the value class, but we'd need to
-295   * convert it to InetSocketAddress at 
some point before the HDFS API call, and
-296   * it seems a bit weird to store 
ServerName since ServerName refers to RegionServers
-297   * and here we really mean DataNode 
locations.
-298   */
-299  protected final MapString, 
InetSocketAddress[] regionFavoredNodesMap =
-300  new ConcurrentHashMap();
-301
-302  // Leases
-303  protected Leases leases;
+264  /**
+265   * Go here to get table descriptors.
+266   */
+267  protected TableDescriptors 
tableDescriptors;
+268
+269  // Replication services. If no 
replication, this handler will be null.
+270  protected ReplicationSourceService 
replicationSourceHandler;
+271  protected ReplicationSinkService 
replicationSinkHandler;
+272
+273  // Compactions
+274  public CompactSplit 
compactSplitThread;
+275
+276  /**
+277   * Map of regions currently being 
served by this region server. Key is the
+278   * encoded region name.  All access 
should be synchronized.
+279   */
+280  protected final MapString, 
HRegion onlineRegions = new ConcurrentHashMap();
+281
+282  /**
+283   * Map of encoded region names to the 
DataNode locations they should be hosted on
+284   * We store the value as 
InetSocketAddress since this is used only in HDFS
+285   * API (create() that takes favored 
nodes as hints for placing file blocks).
+286   * We could have used ServerName here 
as the value class, but we'd need to
+287   * convert it to InetSocketAddress at 
some point before the HDFS API call, and
+288   * it seems a bit weird to store 
ServerName since ServerName refers to RegionServers
+289   * and here we really mean DataNode 
locations.
+290   */
+291  protected final MapString, 
InetSocketAddress[] regionFavoredNodesMap =
+292  new ConcurrentHashMap();
+293
+294  // Leases
+295  protected Leases leases;
+296
+297  // Instance of the hbase executor 
executorService.
+298  protected ExecutorService 
executorService;
+299
+300  // If false, the file system has become 
unavailable
+301  protected volatile boolean fsOk;
+302  protected HFileSystem fs;
+303  protected HFileSystem walFs;
 304
-305  // Instance of the hbase executor 
executorService.
-306  protected ExecutorService 
executorService;
-307
-308  // If false, the file system has become 
unavailable
-309  protected volatile boolean fsOk;
-310  protected HFileSystem fs;
-311  protected HFileSystem walFs;
-312
-313  // Set when a report to the master 
comes back with a message asking us to
-314  // shutdown. Also set by call to stop 
when debugging or running unit tests
-315  // of HRegionServer in isolation.
-316  private volatile boolean stopped = 
false;
-317
-318  // Go down hard. Used if file system 
becomes unavailable and also in
-319  // debugging and unit tests.
-320  private volatile boolean 
abortRequested;
-321  public static final String 
ABORT_TIMEOUT = "hbase.regionserver.abort.timeout";
-322  // Default abort timeout 

[23/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.

2018-11-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html
index cb0eb2d..a08c4ed 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -129,7 +129,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class ModifyNamespaceProcedure
+public class ModifyNamespaceProcedure
 extends AbstractStateMachineNamespaceProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyNamespaceState
 The procedure to add a namespace to an existing table.
 
@@ -191,10 +191,6 @@ extends private NamespaceDescriptor
 oldNsDescriptor
 
-
-private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
-traceEnabled
-
 
 
 
@@ -210,6 +206,13 @@ extends Procedure
 NO_PROC_ID,
 NO_TIMEOUT
 
+
+
+
+
+Fields inherited from 
interfaceorg.apache.hadoop.hbase.master.procedure.TableProcedureInterface
+DUMMY_NAMESPACE_TABLE_NAME
+
 
 
 
@@ -287,67 +290,45 @@ extends 
-private TableNamespaceManager
-getTableNamespaceManager(MasterProcedureEnvenv)
-
-
 TableProcedureInterface.TableOperationType
 getTableOperationType()
 Given an operation type we can take decisions about what to 
do with pending operations.
 
 
-
-private void
-insertIntoNSTable(MasterProcedureEnvenv)
-Insert/update the row into namespace table
-
-
-
+
 protected boolean
 isRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyNamespaceStatestate)
 Used by the default implementation of abort() to know if 
the current state can be aborted
  and rollback can be triggered.
 
 
-
-private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
-isTraceEnabled()
-The procedure could be restarted from a different 
machine.
-
-
-
+
 private boolean
 prepareModify(MasterProcedureEnvenv)
 Action before any real action of adding namespace.
 
 
-
+
 protected void
 rollbackState(MasterProcedureEnvenv,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyNamespaceStatestate)
 called to perform the rollback of the specified state
 
 
-
+
 protected void
 serializeStateData(ProcedureStateSerializerserializer)
 The user-level code of the procedure may have some state to
  persist (e.g.
 
 
-
-private void
-updateZKNamespaceManager(MasterProcedureEnvenv)
-Update ZooKeeper.
-
-
 
 
 
 
 
 Methods inherited from 
classorg.apache.hadoop.hbase.master.procedure.AbstractStateMachineNamespaceProcedure
-acquireLock,
 getTableName,
 releaseLock,
 releaseSyncLatch,
 toStringClassDetails,
 waitInitialized
+acquireLock,
 addOrUpdateNamespace,
 createDirectory,
 createDirectory,
 getTableName,
 getTableNamespaceManager,
 releaseLock,
 releaseSyncLatch,
 toStringClassDetails,
 waitInitialized
 
 
 
@@ -390,7 +371,7 @@ extends 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
@@ -399,25 +380,16 @@ extends 
 
 oldNsDescriptor
-privateNamespaceDescriptor oldNsDescriptor
+privateNamespaceDescriptor oldNsDescriptor
 
 
 
 
 
-
-
-newNsDescriptor
-privateNamespaceDescriptor newNsDescriptor
-
-
-
-
-
 
 
-traceEnabled
-privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
+newNsDescriptor
+privateNamespaceDescriptor newNsDescriptor
 
 
 
@@ -434,7 +406,7 @@ extends 
 
 ModifyNamespaceProcedure
-publicModifyNamespaceProcedure()
+publicModifyNamespaceProcedure()
 
 
 
@@ -443,7 +415,7 @@ extends 
 
 ModifyNamespaceProcedure
-publicModifyNamespaceProcedure(MasterProcedureEnvenv,
+publicModifyNamespaceProcedure(MasterProcedureEnvenv,
 NamespaceDescriptornewNsDescriptor)
 
 
@@ -453,7 +425,7 @@ extends 
 
 ModifyNamespaceProcedure
-publicModifyNamespaceProcedure(MasterProcedureEnvenv,
+publicModifyNamespaceProcedure(MasterProcedureEnvenv,
 NamespaceDescriptornewNsDescriptor,
 

[23/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.

2018-11-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
index 25b7848..5c428b5 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html
@@ -245,20 +245,20 @@
 237return regionEnv != null ? 
regionEnv.getRegion() : null;
 238  }
 239
-240  public TableAuthManager 
getAuthManager() {
+240  public AuthManager getAuthManager() {
 241return 
accessChecker.getAuthManager();
 242  }
 243
 244  private void 
initialize(RegionCoprocessorEnvironment e) throws IOException {
 245final Region region = 
e.getRegion();
 246Configuration conf = 
e.getConfiguration();
-247Mapbyte[], 
ListMultimapString, TablePermission tables = 
AccessControlLists.loadAll(region);
+247Mapbyte[], 
ListMultimapString, UserPermission tables = 
AccessControlLists.loadAll(region);
 248// For each table, write out the 
table's permissions to the respective
 249// znode for that table.
-250for (Map.Entrybyte[], 
ListMultimapString,TablePermission t:
+250for (Map.Entrybyte[], 
ListMultimapString, UserPermission t:
 251  tables.entrySet()) {
 252  byte[] entry = t.getKey();
-253  
ListMultimapString,TablePermission perms = t.getValue();
+253  ListMultimapString, 
UserPermission perms = t.getValue();
 254  byte[] serialized = 
AccessControlLists.writePermissionsAsBytes(perms, conf);
 255  
getAuthManager().getZKPermissionWatcher().writeToZookeeper(entry, 
serialized);
 256}
@@ -294,7 +294,7 @@
 286try (Table t = 
e.getConnection().getTable(AccessControlLists.ACL_TABLE_NAME)) {
 287  for (byte[] entry : entries) {
 288currentEntry = entry;
-289ListMultimapString, 
TablePermission perms =
+289ListMultimapString, 
UserPermission perms =
 290
AccessControlLists.getPermissions(conf, entry, t, null, null, null, false);
 291byte[] serialized = 
AccessControlLists.writePermissionsAsBytes(perms, conf);
 292zkw.writeToZookeeper(entry, 
serialized);
@@ -338,7 +338,7 @@
 330}
 331
 332// 2. check for the table-level, if 
successful we can short-circuit
-333if (getAuthManager().authorize(user, 
tableName, (byte[])null, permRequest)) {
+333if 
(getAuthManager().authorizeUserTable(user, tableName, permRequest)) {
 334  return AuthResult.allow(request, 
"Table permission granted", user,
 335permRequest, tableName, 
families);
 336}
@@ -348,7 +348,7 @@
 340  // all families must pass
 341  for (Map.Entrybyte [], ? 
extends Collection? family : families.entrySet()) {
 342// a) check for family level 
access
-343if 
(getAuthManager().authorize(user, tableName, family.getKey(),
+343if 
(getAuthManager().authorizeUserTable(user, tableName, family.getKey(),
 344permRequest)) {
 345  continue;  // family-level 
permission overrides per-qualifier
 346}
@@ -359,17 +359,17 @@
 351// for each qualifier of the 
family
 352Setbyte[] familySet = 
(Setbyte[])family.getValue();
 353for (byte[] qualifier : 
familySet) {
-354  if 
(!getAuthManager().authorize(user, tableName, family.getKey(),
-355 
qualifier, permRequest)) {
+354  if 
(!getAuthManager().authorizeUserTable(user, tableName,
+355family.getKey(), 
qualifier, permRequest)) {
 356return 
AuthResult.deny(request, "Failed qualifier check", user,
-357permRequest, 
tableName, makeFamilyMap(family.getKey(), qualifier));
+357  permRequest, tableName, 
makeFamilyMap(family.getKey(), qualifier));
 358  }
 359}
 360  } else if (family.getValue() 
instanceof List) { // ListCell
 361ListCell cellList = 
(ListCell)family.getValue();
 362for (Cell cell : cellList) 
{
-363  if 
(!getAuthManager().authorize(user, tableName, family.getKey(),
-364
CellUtil.cloneQualifier(cell), permRequest)) {
+363  if 
(!getAuthManager().authorizeUserTable(user, tableName, family.getKey(),
+364  
CellUtil.cloneQualifier(cell), permRequest)) {
 365return 
AuthResult.deny(request, "Failed qualifier check", user, permRequest,
 366  tableName, 
makeFamilyMap(family.getKey(), CellUtil.cloneQualifier(cell)));
 367  }
@@ -378,7 +378,7 @@
 370} else {
 371  // no qualifiers and 
family-level check already failed

[23/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.

2018-10-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html
index c7d99b2..9d1542c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html
@@ -382,1357 +382,1365 @@
 374for (int i = 0; i  
this.curFunctionCosts.length; i++) {
 375  curFunctionCosts[i] = 
tempFunctionCosts[i];
 376}
-377LOG.info("start 
StochasticLoadBalancer.balancer, initCost=" + currentCost + ", functionCost="
-378+ functionCost());
+377double initCost = currentCost;
+378double newCost = currentCost;
 379
-380double initCost = currentCost;
-381double newCost = currentCost;
-382
-383long computedMaxSteps;
-384if (runMaxSteps) {
-385  computedMaxSteps = 
Math.max(this.maxSteps,
-386  ((long)cluster.numRegions * 
(long)this.stepsPerRegion * (long)cluster.numServers));
-387} else {
-388  computedMaxSteps = 
Math.min(this.maxSteps,
-389  ((long)cluster.numRegions * 
(long)this.stepsPerRegion * (long)cluster.numServers));
-390}
-391// Perform a stochastic walk to see 
if we can get a good fit.
-392long step;
-393
-394for (step = 0; step  
computedMaxSteps; step++) {
-395  Cluster.Action action = 
nextAction(cluster);
-396
-397  if (action.type == Type.NULL) {
-398continue;
-399  }
-400
-401  cluster.doAction(action);
-402  updateCostsWithAction(cluster, 
action);
-403
-404  newCost = computeCost(cluster, 
currentCost);
-405
-406  // Should this be kept?
-407  if (newCost  currentCost) {
-408currentCost = newCost;
-409
-410// save for JMX
-411curOverallCost = currentCost;
-412for (int i = 0; i  
this.curFunctionCosts.length; i++) {
-413  curFunctionCosts[i] = 
tempFunctionCosts[i];
-414}
-415  } else {
-416// Put things back the way they 
were before.
-417// TODO: undo by remembering old 
values
-418Action undoAction = 
action.undoAction();
-419cluster.doAction(undoAction);
-420updateCostsWithAction(cluster, 
undoAction);
-421  }
-422
-423  if 
(EnvironmentEdgeManager.currentTime() - startTime 
-424  maxRunningTime) {
-425break;
-426  }
-427}
-428long endTime = 
EnvironmentEdgeManager.currentTime();
-429
-430
metricsBalancer.balanceCluster(endTime - startTime);
-431
-432// update costs metrics
-433updateStochasticCosts(tableName, 
curOverallCost, curFunctionCosts);
-434if (initCost  currentCost) {
-435  plans = 
createRegionPlans(cluster);
-436  LOG.info("Finished computing new 
load balance plan. Computation took {}" +
-437" to try {} different iterations. 
 Found a solution that moves " +
-438"{} regions; Going from a 
computed cost of {}" +
-439" to a new cost of {}", 
java.time.Duration.ofMillis(endTime - startTime),
-440step, plans.size(), initCost, 
currentCost);
-441  return plans;
-442}
-443LOG.info("Could not find a better 
load balance plan.  Tried {} different configurations in " +
-444  "{}, and did not find anything with 
a computed cost less than {}", step,
-445  java.time.Duration.ofMillis(endTime 
- startTime), initCost);
-446return null;
-447  }
-448
-449  /**
-450   * update costs to JMX
-451   */
-452  private void 
updateStochasticCosts(TableName tableName, Double overall, Double[] subCosts) 
{
-453if (tableName == null) return;
-454
-455// check if the metricsBalancer is 
MetricsStochasticBalancer before casting
-456if (metricsBalancer instanceof 
MetricsStochasticBalancer) {
-457  MetricsStochasticBalancer balancer 
= (MetricsStochasticBalancer) metricsBalancer;
-458  // overall cost
-459  
balancer.updateStochasticCost(tableName.getNameAsString(),
-460"Overall", "Overall cost", 
overall);
-461
-462  // each cost function
-463  for (int i = 0; i  
costFunctions.length; i++) {
-464CostFunction costFunction = 
costFunctions[i];
-465String costFunctionName = 
costFunction.getClass().getSimpleName();
-466Double costPercent = (overall == 
0) ? 0 : (subCosts[i] / overall);
-467// TODO: cost function may need a 
specific description
-468
balancer.updateStochasticCost(tableName.getNameAsString(), costFunctionName,
-469  "The percent of " + 
costFunctionName, costPercent);
-470  }
-471}
-472  }

[23/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.

2018-10-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html
--
diff --git 
a/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html 
b/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html
index 0c894de..8729895 100644
--- a/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html
+++ b/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html
@@ -179,4145 +179,4146 @@
 171 * avoiding port contention if another 
local HBase instance is already running).
 172 * pTo preserve test data 
directories, pass the system property "hbase.testing.preserve.testdir"
 173 * setting it to true.
-174 */
-175@InterfaceAudience.Public
-176@SuppressWarnings("deprecation")
-177public class HBaseTestingUtility extends 
HBaseZKTestingUtility {
-178
-179  /**
-180   * System property key to get test 
directory value. Name is as it is because mini dfs has
-181   * hard-codings to put test data here. 
It should NOT be used directly in HBase, as it's a property
-182   * used in mini dfs.
-183   * @deprecated can be used only with 
mini dfs
-184   */
-185  @Deprecated
-186  private static final String 
TEST_DIRECTORY_KEY = "test.build.data";
-187
-188  public static final String 
REGIONS_PER_SERVER_KEY = "hbase.test.regions-per-server";
-189  /**
-190   * The default number of regions per 
regionserver when creating a pre-split
-191   * table.
-192   */
-193  public static final int 
DEFAULT_REGIONS_PER_SERVER = 3;
-194
+174 * Trigger pre commit.
+175 */
+176@InterfaceAudience.Public
+177@SuppressWarnings("deprecation")
+178public class HBaseTestingUtility extends 
HBaseZKTestingUtility {
+179
+180  /**
+181   * System property key to get test 
directory value. Name is as it is because mini dfs has
+182   * hard-codings to put test data here. 
It should NOT be used directly in HBase, as it's a property
+183   * used in mini dfs.
+184   * @deprecated can be used only with 
mini dfs
+185   */
+186  @Deprecated
+187  private static final String 
TEST_DIRECTORY_KEY = "test.build.data";
+188
+189  public static final String 
REGIONS_PER_SERVER_KEY = "hbase.test.regions-per-server";
+190  /**
+191   * The default number of regions per 
regionserver when creating a pre-split
+192   * table.
+193   */
+194  public static final int 
DEFAULT_REGIONS_PER_SERVER = 3;
 195
-196  public static final String 
PRESPLIT_TEST_TABLE_KEY = "hbase.test.pre-split-table";
-197  public static final boolean 
PRESPLIT_TEST_TABLE = true;
-198
-199  private MiniDFSCluster dfsCluster = 
null;
-200
-201  private volatile HBaseCluster 
hbaseCluster = null;
-202  private MiniMRCluster mrCluster = 
null;
-203
-204  /** If there is a mini cluster running 
for this testing utility instance. */
-205  private volatile boolean 
miniClusterRunning;
-206
-207  private String hadoopLogDir;
-208
-209  /** Directory on test filesystem where 
we put the data for this instance of
-210* HBaseTestingUtility*/
-211  private Path dataTestDirOnTestFS = 
null;
-212
-213  /**
-214   * Shared cluster connection.
-215   */
-216  private volatile Connection 
connection;
-217
-218  /** Filesystem URI used for map-reduce 
mini-cluster setup */
-219  private static String FS_URI;
-220
-221  /** This is for unit tests 
parameterized with a single boolean. */
-222  public static final 
ListObject[] MEMSTORETS_TAGS_PARAMETRIZED = 
memStoreTSAndTagsCombination();
-223
-224  /**
-225   * Checks to see if a specific port is 
available.
-226   *
-227   * @param port the port number to check 
for availability
-228   * @return tttrue/tt if 
the port is available, or ttfalse/tt if not
-229   */
-230  public static boolean available(int 
port) {
-231ServerSocket ss = null;
-232DatagramSocket ds = null;
-233try {
-234  ss = new ServerSocket(port);
-235  ss.setReuseAddress(true);
-236  ds = new DatagramSocket(port);
-237  ds.setReuseAddress(true);
-238  return true;
-239} catch (IOException e) {
-240  // Do nothing
-241} finally {
-242  if (ds != null) {
-243ds.close();
-244  }
-245
-246  if (ss != null) {
-247try {
-248  ss.close();
-249} catch (IOException e) {
-250  /* should not be thrown */
-251}
-252  }
-253}
-254
-255return false;
-256  }
-257
-258  /**
-259   * Create all combinations of Bloom 
filters and compression algorithms for
-260   * testing.
-261   */
-262  private static ListObject[] 
bloomAndCompressionCombinations() {
-263ListObject[] configurations = 
new ArrayList();
-264for (Compression.Algorithm comprAlgo 
:
-265 
HBaseCommonTestingUtility.COMPRESSION_ALGORITHMS) {
-266  for (BloomType bloomType : 
BloomType.values()) {
-267configurations.add(new Object[] { 
comprAlgo, bloomType });
-268  }
-269}
-270return 

[23/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.

2018-10-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html
index 2be5d7b..90d9eed 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9};
+var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -111,7 +111,7 @@ var activeTableTab = "activeTableTab";
 
 @InterfaceAudience.Private
  @InterfaceStability.Evolving
-public final class MasterProcedureUtil
+public final class MasterProcedureUtil
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 
 
@@ -221,6 +221,12 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 toUserInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationuserInfoProto)
 
 
+static https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
+unwrapRemoteIOException(Procedureproc)
+This is a version of unwrapRemoteIOException that can do 
DoNotRetryIOE.
+
+
+
 static boolean
 validateProcedureWALFilename(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringfilename)
 A Procedure WAL file name is of the format: 
pv-wal-id.log where wal-id is 20 digits.
@@ -254,7 +260,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 pattern
-private static finalhttps://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true;
 title="class or interface in java.util.regex">Pattern pattern
+private static finalhttps://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true;
 title="class or interface in java.util.regex">Pattern pattern
 Pattern used to validate a Procedure WAL file name see
  validateProcedureWALFilename(String)
 for description.
 
@@ -273,7 +279,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 MasterProcedureUtil
-privateMasterProcedureUtil()
+privateMasterProcedureUtil()
 
 
 
@@ -290,7 +296,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 toProtoUserInfo
-public 
staticorg.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationtoProtoUserInfo(Useruser)
+public 
staticorg.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationtoProtoUserInfo(Useruser)
 
 
 
@@ -299,7 +305,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 toUserInfo
-public staticUsertoUserInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationuserInfoProto)
+public staticUsertoUserInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationuserInfoProto)
 
 
 
@@ -308,7 +314,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 submitProcedure
-public staticlongsubmitProcedure(MasterProcedureUtil.NonceProcedureRunnablerunnable)
+public staticlongsubmitProcedure(MasterProcedureUtil.NonceProcedureRunnablerunnable)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Helper used to deal with submitting procs with nonce.
  Internally the NonceProcedureRunnable.run() will be called only if no one else
@@ -337,7 +343,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 validateProcedureWALFilename
-public staticbooleanvalidateProcedureWALFilename(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringfilename)
+public staticbooleanvalidateProcedureWALFilename(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringfilename)
 A Procedure WAL file name is of the format: 
pv-wal-id.log where wal-id is 20 digits.
 
 Parameters:
@@ -353,7 +359,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getTablePriority
-public staticintgetTablePriority(TableNametableName)
+public staticintgetTablePriority(TableNametableName)
 Return the priority for the given table. Now meta table is 
3, other system tables are 2, and
  user tables are 1.
 
@@ -361,14 +367,27 

[23/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.

2018-10-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/client/TestMobRestoreSnapshotFromClientClone.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/client/TestMobRestoreSnapshotFromClientClone.html
 
b/testdevapidocs/org/apache/hadoop/hbase/client/TestMobRestoreSnapshotFromClientClone.html
new file mode 100644
index 000..9fd5cd8
--- /dev/null
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/client/TestMobRestoreSnapshotFromClientClone.html
@@ -0,0 +1,423 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+TestMobRestoreSnapshotFromClientClone (Apache HBase 3.0.0-SNAPSHOT Test 
API)
+
+
+
+
+
+var methods = {"i0":10,"i1":10,"i2":9,"i3":9,"i4":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.client
+Class 
TestMobRestoreSnapshotFromClientClone
+
+
+
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase
+
+
+org.apache.hadoop.hbase.client.RestoreSnapshotFromClientCloneTestBase
+
+
+org.apache.hadoop.hbase.client.TestMobRestoreSnapshotFromClientClone
+
+
+
+
+
+
+
+
+
+
+
+
+public class TestMobRestoreSnapshotFromClientClone
+extends RestoreSnapshotFromClientCloneTestBase
+
+
+
+
+
+
+
+
+
+
+
+Field Summary
+
+Fields
+
+Modifier and Type
+Field and Description
+
+
+static HBaseClassTestRule
+CLASS_RULE
+
+
+
+
+
+
+Fields inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase
+admin,
 emptySnapshot,
 FAMILY,
 name,
 snapshot0Rows,
 snapshot1Rows,
 snapshotName0,
 snapshotName1,
 snapshotName2,
 tableName,
 TEST_FAMILY2,
 TEST_UTIL
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+TestMobRestoreSnapshotFromClientClone()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsStatic MethodsInstance MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+protected int
+countRows(org.apache.hadoop.hbase.client.Tabletable,
+ byte[]...families)
+
+
+protected void
+createTable()
+
+
+static void
+setupCluster()
+
+
+protected static void
+setupConf(org.apache.hadoop.conf.Configurationconf)
+
+
+protected void
+verifyRowCount(HBaseTestingUtilityutil,
+  org.apache.hadoop.hbase.TableNametableName,
+  longexpectedRows)
+
+
+
+
+
+
+Methods inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientCloneTestBase
+testCloneAndRestoreSnapshot,
 testCloneSnapshotOfCloned
+
+
+
+
+
+Methods inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase
+getNumReplicas,
 getValidMethodName,
 setup,
 splitRegion,
 tearDown,
 tearDownAfterClass
+
+
+
+
+
+Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
 title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
 title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
 title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/ja
 va/lang/Object.html?is-external=true#notifyAll--" title="class or interface in 

[23/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.

2018-10-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
index 061ce80..bdfc3f8 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
@@ -39,2126 +39,2163 @@
 031import java.util.Set;
 032import 
java.util.concurrent.ConcurrentHashMap;
 033import 
java.util.concurrent.CopyOnWriteArrayList;
-034import java.util.concurrent.TimeUnit;
-035import 
java.util.concurrent.atomic.AtomicBoolean;
-036import 
java.util.concurrent.atomic.AtomicInteger;
-037import 
java.util.concurrent.atomic.AtomicLong;
-038import java.util.stream.Collectors;
-039import java.util.stream.Stream;
-040import 
org.apache.hadoop.conf.Configuration;
-041import 
org.apache.hadoop.hbase.HConstants;
-042import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
-043import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-044import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
-045import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
-046import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
-047import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
-048import 
org.apache.hadoop.hbase.security.User;
-049import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-050import 
org.apache.hadoop.hbase.util.IdLock;
-051import 
org.apache.hadoop.hbase.util.NonceKey;
-052import 
org.apache.hadoop.hbase.util.Threads;
-053import 
org.apache.yetus.audience.InterfaceAudience;
-054import org.slf4j.Logger;
-055import org.slf4j.LoggerFactory;
-056
-057import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-058import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+034import java.util.concurrent.Executor;
+035import java.util.concurrent.Executors;
+036import java.util.concurrent.TimeUnit;
+037import 
java.util.concurrent.atomic.AtomicBoolean;
+038import 
java.util.concurrent.atomic.AtomicInteger;
+039import 
java.util.concurrent.atomic.AtomicLong;
+040import java.util.stream.Collectors;
+041import java.util.stream.Stream;
+042import 
org.apache.hadoop.conf.Configuration;
+043import 
org.apache.hadoop.hbase.HConstants;
+044import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
+045import 
org.apache.hadoop.hbase.log.HBaseMarkers;
+046import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
+047import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
+048import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
+049import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureStoreListener;
+050import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
+051import 
org.apache.hadoop.hbase.security.User;
+052import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+053import 
org.apache.hadoop.hbase.util.IdLock;
+054import 
org.apache.hadoop.hbase.util.NonceKey;
+055import 
org.apache.hadoop.hbase.util.Threads;
+056import 
org.apache.yetus.audience.InterfaceAudience;
+057import org.slf4j.Logger;
+058import org.slf4j.LoggerFactory;
 059
-060import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
-061
-062/**
-063 * Thread Pool that executes the 
submitted procedures.
-064 * The executor has a ProcedureStore 
associated.
-065 * Each operation is logged and on 
restart the pending procedures are resumed.
-066 *
-067 * Unless the Procedure code throws an 
error (e.g. invalid user input)
-068 * the procedure will complete (at some 
point in time), On restart the pending
-069 * procedures are resumed and the once 
failed will be rolledback.
+060import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+061import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+062import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
+063
+064import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
+065
+066/**
+067 * Thread Pool that executes the 
submitted procedures.
+068 * The executor has a ProcedureStore 
associated.
+069 * Each operation is logged and on 
restart the pending procedures are resumed.
 070 *
-071 * The user can add procedures to the 
executor via submitProcedure(proc)
-072 * check for the finished state via 
isFinished(procId)
-073 * and get the result via 
getResult(procId)
-074 */
-075@InterfaceAudience.Private
-076public class 
ProcedureExecutorTEnvironment {
-077  private static final Logger LOG = 
LoggerFactory.getLogger(ProcedureExecutor.class);
-078
-079  public static final String 

[23/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.

2018-09-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
index 43c66a8..061ce80 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
@@ -23,2136 +23,2142 @@
 015 * See the License for the specific 
language governing permissions and
 016 * limitations under the License.
 017 */
-018
-019package 
org.apache.hadoop.hbase.procedure2;
-020
-021import java.io.IOException;
-022import java.util.ArrayDeque;
-023import java.util.ArrayList;
-024import java.util.Arrays;
-025import java.util.Collection;
-026import java.util.Deque;
-027import java.util.HashSet;
-028import java.util.Iterator;
-029import java.util.List;
-030import java.util.Map;
-031import java.util.Objects;
-032import java.util.Set;
-033import 
java.util.concurrent.ConcurrentHashMap;
-034import 
java.util.concurrent.CopyOnWriteArrayList;
-035import java.util.concurrent.TimeUnit;
-036import 
java.util.concurrent.atomic.AtomicBoolean;
-037import 
java.util.concurrent.atomic.AtomicInteger;
-038import 
java.util.concurrent.atomic.AtomicLong;
-039import java.util.stream.Collectors;
-040import java.util.stream.Stream;
-041
-042import 
org.apache.hadoop.conf.Configuration;
-043import 
org.apache.hadoop.hbase.HConstants;
-044import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
-045import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-046import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
-047import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
-048import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
-049import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
-050import 
org.apache.hadoop.hbase.security.User;
-051import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-052import 
org.apache.hadoop.hbase.util.IdLock;
-053import 
org.apache.hadoop.hbase.util.NonceKey;
-054import 
org.apache.hadoop.hbase.util.Threads;
-055import 
org.apache.yetus.audience.InterfaceAudience;
-056import org.slf4j.Logger;
-057import org.slf4j.LoggerFactory;
-058
-059import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-060import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+018package 
org.apache.hadoop.hbase.procedure2;
+019
+020import java.io.IOException;
+021import java.util.ArrayDeque;
+022import java.util.ArrayList;
+023import java.util.Arrays;
+024import java.util.Collection;
+025import java.util.Deque;
+026import java.util.HashSet;
+027import java.util.Iterator;
+028import java.util.List;
+029import java.util.Map;
+030import java.util.Objects;
+031import java.util.Set;
+032import 
java.util.concurrent.ConcurrentHashMap;
+033import 
java.util.concurrent.CopyOnWriteArrayList;
+034import java.util.concurrent.TimeUnit;
+035import 
java.util.concurrent.atomic.AtomicBoolean;
+036import 
java.util.concurrent.atomic.AtomicInteger;
+037import 
java.util.concurrent.atomic.AtomicLong;
+038import java.util.stream.Collectors;
+039import java.util.stream.Stream;
+040import 
org.apache.hadoop.conf.Configuration;
+041import 
org.apache.hadoop.hbase.HConstants;
+042import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
+043import 
org.apache.hadoop.hbase.log.HBaseMarkers;
+044import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
+045import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
+046import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
+047import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
+048import 
org.apache.hadoop.hbase.security.User;
+049import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+050import 
org.apache.hadoop.hbase.util.IdLock;
+051import 
org.apache.hadoop.hbase.util.NonceKey;
+052import 
org.apache.hadoop.hbase.util.Threads;
+053import 
org.apache.yetus.audience.InterfaceAudience;
+054import org.slf4j.Logger;
+055import org.slf4j.LoggerFactory;
+056
+057import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+058import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+059
+060import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
 061
-062import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
-063
-064/**
-065 * Thread Pool that executes the 
submitted procedures.
-066 * The executor has a ProcedureStore 
associated.
-067 * Each operation is logged and on 
restart the pending procedures 

[23/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.

2018-09-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html 
b/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html
index aa153a2..5aa7403 100644
--- a/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html
+++ b/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public static class WALSplitter.WriterThread
+public static class WALSplitter.WriterThread
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true;
 title="class or interface in java.lang">Thread
 
 
@@ -266,7 +266,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 shouldStop
-private volatileboolean shouldStop
+private volatileboolean shouldStop
 
 
 
@@ -275,7 +275,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 controller
-privateWALSplitter.PipelineController controller
+privateWALSplitter.PipelineController controller
 
 
 
@@ -284,7 +284,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 entryBuffers
-privateWALSplitter.EntryBuffers entryBuffers
+privateWALSplitter.EntryBuffers entryBuffers
 
 
 
@@ -293,7 +293,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 outputSink
-privateWALSplitter.OutputSink outputSink
+privateWALSplitter.OutputSink outputSink
 
 
 
@@ -310,7 +310,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 WriterThread
-WriterThread(WALSplitter.PipelineControllercontroller,
+WriterThread(WALSplitter.PipelineControllercontroller,
  WALSplitter.EntryBuffersentryBuffers,
  WALSplitter.OutputSinksink,
  inti)
@@ -330,7 +330,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 run
-publicvoidrun()
+publicvoidrun()
 
 Specified by:
 https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--;
 title="class or interface in java.lang">runin 
interfacehttps://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true;
 title="class or interface in java.lang">Runnable
@@ -345,7 +345,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 doRun
-privatevoiddoRun()
+privatevoiddoRun()
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
 Throws:
@@ -359,7 +359,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 writeBuffer
-privatevoidwriteBuffer(WALSplitter.RegionEntryBufferbuffer)
+privatevoidwriteBuffer(WALSplitter.RegionEntryBufferbuffer)
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
 Throws:
@@ -373,7 +373,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 finish
-voidfinish()
+voidfinish()
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html 
b/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html
index 0257e4b..621546b 100644
--- a/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html
+++ b/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html
@@ -228,75 +228,75 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 fileBeingSplit
 
 
-protected 
org.apache.hadoop.fs.FileSystem
-fs
-
-
 protected https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long
 lastFlushedSequenceIds
 
-
+
 private static org.slf4j.Logger
 LOG
 
-
+
 private static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 OLD_SEQUENCE_ID_FILE_SUFFIX
 
-
+
 (package private) WALSplitter.OutputSink
 outputSink
 
-
+
 private static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 RECOVERED_LOG_TMPFILE_SUFFIX
 
-
+
 protected https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in 

[23/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.

2018-09-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html
index 0cf012a..976894f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html
@@ -63,3884 +63,3883 @@
 055import javax.servlet.http.HttpServlet;
 056import 
javax.servlet.http.HttpServletRequest;
 057import 
javax.servlet.http.HttpServletResponse;
-058
-059import 
org.apache.commons.lang3.StringUtils;
-060import 
org.apache.hadoop.conf.Configuration;
-061import org.apache.hadoop.fs.Path;
-062import 
org.apache.hadoop.hbase.ChoreService;
-063import 
org.apache.hadoop.hbase.ClusterId;
-064import 
org.apache.hadoop.hbase.ClusterMetrics;
-065import 
org.apache.hadoop.hbase.ClusterMetrics.Option;
-066import 
org.apache.hadoop.hbase.ClusterMetricsBuilder;
-067import 
org.apache.hadoop.hbase.CompoundConfiguration;
-068import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-069import 
org.apache.hadoop.hbase.HBaseIOException;
-070import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-071import 
org.apache.hadoop.hbase.HConstants;
-072import 
org.apache.hadoop.hbase.InvalidFamilyOperationException;
-073import 
org.apache.hadoop.hbase.MasterNotRunningException;
-074import 
org.apache.hadoop.hbase.MetaTableAccessor;
-075import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-076import 
org.apache.hadoop.hbase.PleaseHoldException;
-077import 
org.apache.hadoop.hbase.ReplicationPeerNotFoundException;
-078import 
org.apache.hadoop.hbase.ServerName;
-079import 
org.apache.hadoop.hbase.TableDescriptors;
-080import 
org.apache.hadoop.hbase.TableName;
-081import 
org.apache.hadoop.hbase.TableNotDisabledException;
-082import 
org.apache.hadoop.hbase.TableNotFoundException;
-083import 
org.apache.hadoop.hbase.UnknownRegionException;
-084import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
-085import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
-086import 
org.apache.hadoop.hbase.client.MasterSwitchType;
-087import 
org.apache.hadoop.hbase.client.RegionInfo;
-088import 
org.apache.hadoop.hbase.client.RegionInfoBuilder;
-089import 
org.apache.hadoop.hbase.client.Result;
-090import 
org.apache.hadoop.hbase.client.TableDescriptor;
-091import 
org.apache.hadoop.hbase.client.TableDescriptorBuilder;
-092import 
org.apache.hadoop.hbase.client.TableState;
-093import 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
-094import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-095import 
org.apache.hadoop.hbase.exceptions.MergeRegionException;
-096import 
org.apache.hadoop.hbase.executor.ExecutorType;
-097import 
org.apache.hadoop.hbase.favored.FavoredNodesManager;
-098import 
org.apache.hadoop.hbase.favored.FavoredNodesPromoter;
-099import 
org.apache.hadoop.hbase.http.InfoServer;
-100import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
-101import 
org.apache.hadoop.hbase.ipc.RpcServer;
-102import 
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
-103import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-104import 
org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode;
-105import 
org.apache.hadoop.hbase.master.assignment.AssignProcedure;
-106import 
org.apache.hadoop.hbase.master.assignment.AssignmentManager;
-107import 
org.apache.hadoop.hbase.master.assignment.MergeTableRegionsProcedure;
-108import 
org.apache.hadoop.hbase.master.assignment.MoveRegionProcedure;
-109import 
org.apache.hadoop.hbase.master.assignment.RegionStateNode;
-110import 
org.apache.hadoop.hbase.master.assignment.RegionStates;
-111import 
org.apache.hadoop.hbase.master.assignment.TransitRegionStateProcedure;
-112import 
org.apache.hadoop.hbase.master.assignment.UnassignProcedure;
-113import 
org.apache.hadoop.hbase.master.balancer.BalancerChore;
-114import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer;
-115import 
org.apache.hadoop.hbase.master.balancer.ClusterStatusChore;
-116import 
org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory;
-117import 
org.apache.hadoop.hbase.master.cleaner.CleanerChore;
-118import 
org.apache.hadoop.hbase.master.cleaner.HFileCleaner;
-119import 
org.apache.hadoop.hbase.master.cleaner.LogCleaner;
-120import 
org.apache.hadoop.hbase.master.cleaner.ReplicationBarrierCleaner;
-121import 
org.apache.hadoop.hbase.master.locking.LockManager;
-122import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
-123import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType;
-124import 
org.apache.hadoop.hbase.master.normalizer.RegionNormalizer;
-125import 

[23/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.

2018-09-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html
index a5789e0..93a57cb 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html
@@ -238,4120 +238,4119 @@
 230 * @see Admin
 231 */
 232@InterfaceAudience.Private
-233@InterfaceStability.Evolving
-234public class HBaseAdmin implements Admin 
{
-235  private static final Logger LOG = 
LoggerFactory.getLogger(HBaseAdmin.class);
-236
-237  private ClusterConnection connection;
-238
-239  private final Configuration conf;
-240  private final long pause;
-241  private final int numRetries;
-242  private final int syncWaitTimeout;
-243  private boolean aborted;
-244  private int operationTimeout;
-245  private int rpcTimeout;
-246
-247  private RpcRetryingCallerFactory 
rpcCallerFactory;
-248  private RpcControllerFactory 
rpcControllerFactory;
-249
-250  private NonceGenerator ng;
-251
-252  @Override
-253  public int getOperationTimeout() {
-254return operationTimeout;
-255  }
-256
-257  HBaseAdmin(ClusterConnection 
connection) throws IOException {
-258this.conf = 
connection.getConfiguration();
-259this.connection = connection;
-260
-261// TODO: receive 
ConnectionConfiguration here rather than re-parsing these configs every time.
-262this.pause = 
this.conf.getLong(HConstants.HBASE_CLIENT_PAUSE,
-263
HConstants.DEFAULT_HBASE_CLIENT_PAUSE);
-264this.numRetries = 
this.conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER,
-265
HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER);
-266this.operationTimeout = 
this.conf.getInt(HConstants.HBASE_CLIENT_OPERATION_TIMEOUT,
-267
HConstants.DEFAULT_HBASE_CLIENT_OPERATION_TIMEOUT);
-268this.rpcTimeout = 
this.conf.getInt(HConstants.HBASE_RPC_TIMEOUT_KEY,
-269
HConstants.DEFAULT_HBASE_RPC_TIMEOUT);
-270this.syncWaitTimeout = 
this.conf.getInt(
-271  
"hbase.client.sync.wait.timeout.msec", 10 * 6); // 10min
-272
-273this.rpcCallerFactory = 
connection.getRpcRetryingCallerFactory();
-274this.rpcControllerFactory = 
connection.getRpcControllerFactory();
-275
-276this.ng = 
this.connection.getNonceGenerator();
-277  }
-278
-279  @Override
-280  public void abort(String why, Throwable 
e) {
-281// Currently does nothing but throw 
the passed message and exception
-282this.aborted = true;
-283throw new RuntimeException(why, e);
-284  }
-285
-286  @Override
-287  public boolean isAborted() {
-288return this.aborted;
-289  }
-290
-291  @Override
-292  public boolean abortProcedure(final 
long procId, final boolean mayInterruptIfRunning)
-293  throws IOException {
-294return 
get(abortProcedureAsync(procId, mayInterruptIfRunning), this.syncWaitTimeout,
-295  TimeUnit.MILLISECONDS);
-296  }
-297
-298  @Override
-299  public FutureBoolean 
abortProcedureAsync(final long procId, final boolean mayInterruptIfRunning)
-300  throws IOException {
-301Boolean abortProcResponse =
-302executeCallable(new 
MasterCallableAbortProcedureResponse(getConnection(),
-303getRpcControllerFactory()) 
{
-304  @Override
-305  protected AbortProcedureResponse 
rpcCall() throws Exception {
-306AbortProcedureRequest 
abortProcRequest =
-307
AbortProcedureRequest.newBuilder().setProcId(procId).build();
-308return 
master.abortProcedure(getRpcController(), abortProcRequest);
-309  }
-310}).getIsProcedureAborted();
-311return new AbortProcedureFuture(this, 
procId, abortProcResponse);
-312  }
-313
-314  @Override
-315  public ListTableDescriptor 
listTableDescriptors() throws IOException {
-316return 
listTableDescriptors((Pattern)null, false);
-317  }
-318
-319  @Override
-320  public ListTableDescriptor 
listTableDescriptors(Pattern pattern) throws IOException {
-321return listTableDescriptors(pattern, 
false);
-322  }
-323
-324  @Override
-325  public ListTableDescriptor 
listTableDescriptors(Pattern pattern, boolean includeSysTables)
-326  throws IOException {
-327return executeCallable(new 
MasterCallableListTableDescriptor(getConnection(),
-328getRpcControllerFactory()) {
-329  @Override
-330  protected 
ListTableDescriptor rpcCall() throws Exception {
-331GetTableDescriptorsRequest req 
=
-332
RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables);
-333return 

[23/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.

2018-09-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
index 309387e..11a1dff 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
@@ -27,211 +27,206 @@
 019
 020import java.io.IOException;
 021import java.util.ArrayList;
-022import java.util.Arrays;
-023import java.util.Comparator;
-024import java.util.Objects;
-025import java.util.TreeSet;
-026
-027import org.apache.hadoop.hbase.Cell;
-028import 
org.apache.hadoop.hbase.CellUtil;
-029import 
org.apache.hadoop.hbase.PrivateCellUtil;
-030import 
org.apache.yetus.audience.InterfaceAudience;
-031import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-032import 
org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException;
-033import 
org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations;
-034import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-035import 
org.apache.hadoop.hbase.util.Bytes;
-036
-037/**
-038 * This filter is used for selecting only 
those keys with columns that matches
-039 * a particular prefix. For example, if 
prefix is 'an', it will pass keys will
-040 * columns like 'and', 'anti' but not 
keys with columns like 'ball', 'act'.
-041 */
-042@InterfaceAudience.Public
-043public class MultipleColumnPrefixFilter 
extends FilterBase {
-044  protected byte [] hint = null;
-045  protected TreeSetbyte [] 
sortedPrefixes = createTreeSet();
-046  private final static int 
MAX_LOG_PREFIXES = 5;
-047
-048  public MultipleColumnPrefixFilter(final 
byte [][] prefixes) {
-049if (prefixes != null) {
-050  for (int i = 0; i  
prefixes.length; i++) {
-051if 
(!sortedPrefixes.add(prefixes[i]))
-052  throw new 
IllegalArgumentException ("prefixes must be distinct");
-053  }
-054}
-055  }
-056
-057  public byte [][] getPrefix() {
-058int count = 0;
-059byte [][] temp = new byte 
[sortedPrefixes.size()][];
-060for (byte [] prefixes : 
sortedPrefixes) {
-061  temp [count++] = prefixes;
-062}
-063return temp;
-064  }
-065
-066  @Override
-067  public boolean filterRowKey(Cell cell) 
throws IOException {
-068// Impl in FilterBase might do 
unnecessary copy for Off heap backed Cells.
-069return false;
-070  }
-071
-072  @Deprecated
-073  @Override
-074  public ReturnCode filterKeyValue(final 
Cell c) {
-075return filterCell(c);
-076  }
-077
-078  @Override
-079  public ReturnCode filterCell(final Cell 
c) {
-080if (sortedPrefixes.isEmpty()) {
-081  return ReturnCode.INCLUDE;
-082} else {
-083  return filterColumn(c);
-084}
-085  }
-086
-087  public ReturnCode filterColumn(Cell 
cell) {
-088byte [] qualifier = 
CellUtil.cloneQualifier(cell);
-089TreeSetbyte [] 
lesserOrEqualPrefixes =
-090  (TreeSetbyte []) 
sortedPrefixes.headSet(qualifier, true);
-091
-092if (lesserOrEqualPrefixes.size() != 
0) {
-093  byte [] 
largestPrefixSmallerThanQualifier = lesserOrEqualPrefixes.last();
-094  
-095  if (Bytes.startsWith(qualifier, 
largestPrefixSmallerThanQualifier)) {
-096return ReturnCode.INCLUDE;
-097  }
-098  
-099  if (lesserOrEqualPrefixes.size() == 
sortedPrefixes.size()) {
-100return ReturnCode.NEXT_ROW;
-101  } else {
-102hint = 
sortedPrefixes.higher(largestPrefixSmallerThanQualifier);
-103return 
ReturnCode.SEEK_NEXT_USING_HINT;
-104  }
-105} else {
-106  hint = sortedPrefixes.first();
-107  return 
ReturnCode.SEEK_NEXT_USING_HINT;
-108}
-109  }
-110
-111  public static Filter 
createFilterFromArguments(ArrayListbyte [] filterArguments) {
-112byte [][] prefixes = new byte 
[filterArguments.size()][];
-113for (int i = 0 ; i  
filterArguments.size(); i++) {
-114  byte [] columnPrefix = 
ParseFilter.removeQuotesFromByteArray(filterArguments.get(i));
-115  prefixes[i] = columnPrefix;
-116}
-117return new 
MultipleColumnPrefixFilter(prefixes);
-118  }
-119
-120  /**
-121   * @return The filter serialized using 
pb
-122   */
-123  @Override
-124  public byte [] toByteArray() {
-125
FilterProtos.MultipleColumnPrefixFilter.Builder builder =
-126  
FilterProtos.MultipleColumnPrefixFilter.newBuilder();
-127for (byte [] element : 
sortedPrefixes) {
-128  if (element != null) 
builder.addSortedPrefixes(UnsafeByteOperations.unsafeWrap(element));
-129}
-130return 
builder.build().toByteArray();
-131  }
-132
-133  /**
-134   * @param pbBytes A pb serialized 
{@link MultipleColumnPrefixFilter} instance
-135   * 

[23/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.

2018-08-31 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html
index ad86b01..57c6c7e 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = {"i0":10,"i1":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -136,7 +136,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 private MetricsTableWrapperAggregate
-tableWrapperAgg
+wrapper
 
 
 
@@ -178,6 +178,64 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 MetricsTableWrapperAggregate
 getTableWrapperAgg()
 
+
+void
+incrSplitRequest(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringtable)
+
+
+void
+incrSplitSuccess(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringtable)
+
+
+void
+updateCompactionInputFileCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+  booleanisMajor,
+  longc)
+
+
+void
+updateCompactionInputSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+ booleanisMajor,
+ longbytes)
+
+
+void
+updateCompactionOutputFileCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+   booleanisMajor,
+   longc)
+
+
+void
+updateCompactionOutputSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+  booleanisMajor,
+  longbytes)
+
+
+void
+updateCompactionTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+booleanisMajor,
+longt)
+
+
+void
+updateFlushMemstoreSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+   longbytes)
+
+
+void
+updateFlushOutputSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+ longbytes)
+
+
+void
+updateFlushTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+   longt)
+
+
+void
+updateSplitTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+   longt)
+
 
 
 
@@ -209,13 +267,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 private finalMetricsTableAggregateSource tableSourceAgg
 
 
-
+
 
 
 
 
-tableWrapperAgg
-privateMetricsTableWrapperAggregate tableWrapperAgg
+wrapper
+privateMetricsTableWrapperAggregate wrapper
 
 
 
@@ -255,12 +313,125 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 
-
+
 
 getTableSourceAgg
 publicMetricsTableAggregateSourcegetTableSourceAgg()
 
 
+
+
+
+
+
+incrSplitRequest
+publicvoidincrSplitRequest(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable)
+
+
+
+
+
+
+
+incrSplitSuccess
+publicvoidincrSplitSuccess(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable)
+
+
+
+
+
+
+
+updateSplitTime
+publicvoidupdateSplitTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+longt)
+
+
+
+
+
+
+
+updateFlushTime
+publicvoidupdateFlushTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringtable,
+longt)
+
+
+
+
+
+
+
+updateFlushMemstoreSize

[23/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.

2018-08-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
index c9128a4..c5b8dbd 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
@@ -63,750 +63,734 @@
 055public class RegionStates {
 056  private static final Logger LOG = 
LoggerFactory.getLogger(RegionStates.class);
 057
-058  // TODO: need to be more specific, i.e, 
OPENING vs. OPEN, CLOSING vs. CLOSED.
-059  static final State[] 
STATES_EXPECTED_ON_OPEN = new State[] {
-060State.OPEN, // State may already be 
OPEN if we died after receiving the OPEN from regionserver
-061// but before complete 
finish of AssignProcedure. HBASE-20100.
-062State.OFFLINE, State.CLOSED, 
State.ABNORMALLY_CLOSED, // disable/offline
-063State.SPLITTING, // 
ServerCrashProcedure
-064State.OPENING, State.FAILED_OPEN, // 
already in-progress (retrying)
-065State.MERGED, State.SPLITTING_NEW
-066  };
-067
-068  static final State[] 
STATES_EXPECTED_ON_CLOSE = new State[] {
-069State.SPLITTING, State.MERGING, 
State.OPENING, // ServerCrashProcedure
-070State.OPEN,   // 
enabled/open
-071State.CLOSING // 
already in-progress (retrying)
-072  };
-073
-074  // This comparator sorts the 
RegionStates by time stamp then Region name.
-075  // Comparing by timestamp alone can 
lead us to discard different RegionStates that happen
-076  // to share a timestamp.
-077  private static class 
RegionStateStampComparator implements ComparatorRegionState {
-078@Override
-079public int compare(final RegionState 
l, final RegionState r) {
-080  int stampCmp = 
Long.compare(l.getStamp(), r.getStamp());
-081  return stampCmp != 0 ? stampCmp : 
RegionInfo.COMPARATOR.compare(l.getRegion(), r.getRegion());
-082}
-083  }
-084
-085  public final static 
RegionStateStampComparator REGION_STATE_STAMP_COMPARATOR =
-086  new RegionStateStampComparator();
-087
-088  // TODO: Replace the 
ConcurrentSkipListMaps
-089  /**
-090   * RegionName -- i.e. 
RegionInfo.getRegionName() -- as bytes to {@link RegionStateNode}
-091   */
-092  private final 
ConcurrentSkipListMapbyte[], RegionStateNode regionsMap =
-093  new 
ConcurrentSkipListMapbyte[], RegionStateNode(Bytes.BYTES_COMPARATOR);
+058  // This comparator sorts the 
RegionStates by time stamp then Region name.
+059  // Comparing by timestamp alone can 
lead us to discard different RegionStates that happen
+060  // to share a timestamp.
+061  private static class 
RegionStateStampComparator implements ComparatorRegionState {
+062@Override
+063public int compare(final RegionState 
l, final RegionState r) {
+064  int stampCmp = 
Long.compare(l.getStamp(), r.getStamp());
+065  return stampCmp != 0 ? stampCmp : 
RegionInfo.COMPARATOR.compare(l.getRegion(), r.getRegion());
+066}
+067  }
+068
+069  public final static 
RegionStateStampComparator REGION_STATE_STAMP_COMPARATOR =
+070  new RegionStateStampComparator();
+071
+072  // TODO: Replace the 
ConcurrentSkipListMaps
+073  /**
+074   * RegionName -- i.e. 
RegionInfo.getRegionName() -- as bytes to {@link RegionStateNode}
+075   */
+076  private final 
ConcurrentSkipListMapbyte[], RegionStateNode regionsMap =
+077  new 
ConcurrentSkipListMapbyte[], RegionStateNode(Bytes.BYTES_COMPARATOR);
+078
+079  private final 
ConcurrentSkipListMapRegionInfo, RegionStateNode regionInTransition =
+080new 
ConcurrentSkipListMapRegionInfo, 
RegionStateNode(RegionInfo.COMPARATOR);
+081
+082  /**
+083   * Regions marked as offline on a read 
of hbase:meta. Unused or at least, once
+084   * offlined, regions have no means of 
coming on line again. TODO.
+085   */
+086  private final 
ConcurrentSkipListMapRegionInfo, RegionStateNode regionOffline =
+087new 
ConcurrentSkipListMapRegionInfo, RegionStateNode();
+088
+089  private final 
ConcurrentSkipListMapbyte[], RegionFailedOpen regionFailedOpen =
+090new ConcurrentSkipListMapbyte[], 
RegionFailedOpen(Bytes.BYTES_COMPARATOR);
+091
+092  private final 
ConcurrentHashMapServerName, ServerStateNode serverMap =
+093  new 
ConcurrentHashMapServerName, ServerStateNode();
 094
-095  private final 
ConcurrentSkipListMapRegionInfo, RegionStateNode regionInTransition =
-096new 
ConcurrentSkipListMapRegionInfo, 
RegionStateNode(RegionInfo.COMPARATOR);
-097
-098  /**
-099   * Regions marked as offline on a read 
of hbase:meta. Unused or at least, once
-100   * offlined, regions have no means of 
coming on line again. TODO.
-101   */
-102  private final 

[23/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.

2018-08-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/checkstyle-aggregate.html
--
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index 6098f6a..236f443 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Checkstyle Results
 
@@ -9838,12 +9838,12 @@
 http://checkstyle.sourceforge.net/config_javadoc.html#JavadocTagContinuationIndentation;>JavadocTagContinuationIndentation
 
 offset: 2
-763
+764
 Error
 
 
 http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription;>NonEmptyAtclauseDescription
-3606
+3605
 Error
 
 misc
@@ -19479,7 +19479,7 @@
 
 Error
 javadoc
-NonEmptyAtclauseDescription
+JavadocTagContinuationIndentation
 Javadoc comment at column 0 has parse error. Details: no viable 
alternative at input '   *' while parsing JAVADOC_TAG
 117
 
@@ -23225,7 +23225,7 @@
 imports
 ImportOrder
 Wrong order for 
'org.apache.hadoop.hbase.exceptions.DeserializationException' import.
-24
+25
 
 org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java
 
@@ -37613,55 +37613,55 @@
 imports
 ImportOrder
 Wrong order for 
'org.apache.hadoop.hbase.exceptions.DeserializationException' import.
-27
+28
 
 Error
 imports
 ImportOrder
 Wrong order for 
'org.apache.hbase.thirdparty.com.google.common.base.Preconditions' import.
-30
+31
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-56
+57
 
 Error
 annotation
 MissingDeprecated
 Must include both @java.lang.Deprecated annotation and @deprecated Javadoc 
tag with description.
-65
+66
 
 Error
 javadoc
 NonEmptyAtclauseDescription
 At-clause should have a non-empty description.
-103
+104
 
 Error
 indentation
 Indentation
 'throws' has incorrect indentation level 2, expected level should be 
4.
-107
+108
 
 Error
 javadoc
 JavadocTagContinuationIndentation
 Line continuation have incorrect indentation level, expected level should 
be 2.
-120
+121
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-124
+125
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-125
+126
 
 org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java
 
@@ -37676,79 +37676,79 @@
 imports
 ImportOrder
 Wrong order for 
'org.apache.hadoop.hbase.exceptions.DeserializationException' import.
-28
+29
 
 Error
 imports
 ImportOrder
 Wrong order for 'org.apache.hadoop.hbase.util.Bytes' import.
-30
+31
 
 Error
 sizes
 LineLength
 Line is longer than 100 characters (found 115).
-38
+39
 
 Error
 sizes
 LineLength
 Line is longer than 100 characters (found 110).
-39
+40
 
 Error
 blocks
 LeftCurly
 '{' at column 3 should be on the previous line.
-60
+61
 
 Error
 annotation
 MissingDeprecated
 Must include both @java.lang.Deprecated annotation and @deprecated Javadoc 
tag with description.
-113
+114
 
 Error
 blocks
 LeftCurly
 '{' at column 3 should be on the previous line.
-121
+122
 
 Error
 blocks
 LeftCurly
 '{' at column 3 should be on the previous line.
-156
+157
 
 Error
 javadoc
 NonEmptyAtclauseDescription
 At-clause should have a non-empty description.
-188
+189
 
 Error
 indentation
 Indentation
 'throws' has incorrect indentation level 2, expected level should be 
4.
-192
+193
 
 Error
 javadoc
 JavadocTagContinuationIndentation
 Line continuation have incorrect indentation level, expected level should 
be 2.
-209
+210
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-213
+214
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-214
+215
 
 org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java
 
@@ -37763,73 +37763,73 @@
 imports
 ImportOrder
 Wrong order for 
'org.apache.hadoop.hbase.exceptions.DeserializationException' import.
-29
+30
 
 Error
 imports
 ImportOrder
 Wrong order for 'org.apache.hadoop.hbase.util.ByteBufferUtils' import.
-31
+32
 
 Error
 annotation
 MissingDeprecated
 Must include both @java.lang.Deprecated annotation and @deprecated Javadoc 
tag with description.
-61
+62
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-120
+121
 
 Error
 javadoc
 NonEmptyAtclauseDescription
 At-clause should have a non-empty description.
-127
+128
 
 Error
 indentation
 Indentation
 'throws' has incorrect indentation level 2, expected level should be 
4.
-131
+132
 
 Error
 javadoc
 JavadocTagContinuationIndentation
 Line continuation have incorrect indentation level, expected level should 
be 2.
-144
+145
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-148
+149
 
 Error
 indentation
 Indentation
 'if' has incorrect indentation level 3, expected level should be 4.
-148
+149
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-149
+150
 
 Error
 indentation
 Indentation
 'if' has incorrect indentation level 3, expected level should be 4.
-149
+150
 
 Error
 indentation
 Indentation
 'method def' child has incorrect indentation level 3, expected level 
should be 4.
-151
+152
 
 

[23/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  CollectionHStore stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at 

[23/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.

2018-08-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html
index 617a7a2..db068f2 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":42};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -49,8 +49,8 @@ var activeTableTab = "activeTableTab";
 
 
 
-PrevClass
-NextClass
+PrevClass
+NextClass
 
 
 Frames
@@ -122,32 +122,19 @@ var activeTableTab = "activeTableTab";
 https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableProcedureMasterProcedureEnv, TableProcedureInterface, RemoteProcedureDispatcher.RemoteProcedureMasterProcedureEnv,ServerName
 
 
+Deprecated.
+Do not use any 
more.
+
 
-@InterfaceAudience.Private
-public class UnassignProcedure
+https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true;
 title="class or interface in java.lang">@Deprecated
+ @InterfaceAudience.Private
+public class UnassignProcedure
 extends RegionTransitionProcedure
-Procedure that describes the unassignment of a single 
region.
- There can only be one RegionTransitionProcedure -- i.e. an assign or an 
unassign -- per region
- running at a time, since each procedure takes a lock on the region.
-
- The Unassign starts by placing a "close region" request in the Remote 
Dispatcher
- queue, and the procedure will then go into a "waiting state" (suspend).
- The Remote Dispatcher will batch the various requests for that server and
- they will be sent to the RS for execution.
- The RS will complete the open operation by calling 
master.reportRegionStateTransition().
- The AM will intercept the transition report, and notify this procedure.
- The procedure will wakeup and finish the unassign by publishing its new state 
on meta.
- If we are unable to contact the remote regionserver whether because of 
ConnectException
- or socket timeout, we will call expire on the server we were trying to 
contact. We will remain
- in suspended state waiting for a wake up from the ServerCrashProcedure that 
is processing the
- failed server. The basic idea is that if we notice a crashed server, then we 
have a
- responsibility; i.e. we should not let go of the region until we are sure the 
server that was
- hosting has had its crash processed. If we let go of the region before then, 
an assign might
- run before the logs have been split which would make for data loss.
-
- TODO: Rather than this tricky coordination between SCP and this Procedure, 
instead, work on
- returning a SCP as our subprocedure; probably needs work on the framework to 
do this,
- especially if the SCP already created.
+Leave here only for checking if we can successfully start 
the master.
+
+See Also:
+TransitRegionStateProcedure
+
 
 
 
@@ -191,28 +178,26 @@ extends 
 protected ServerName
 destinationServer
-The Server we will subsequently assign the region too (can 
be null).
-
+Deprecated.
+
 
 
 private boolean
-force
+force
+Deprecated.
+
 
 
 protected ServerName
 hostingServer
-Where to send the unassign RPC.
-
+Deprecated.
+
 
 
-private static org.slf4j.Logger
-LOG
-
-
 private boolean
 removeAfterUnassigning
-Whether deleting the region from in-memory states after 
unassigning the region.
-
+Deprecated.
+
 
 
 
@@ -243,26 +228,9 @@ extends Constructor and Description
 
 
-UnassignProcedure()
-
-
-UnassignProcedure(RegionInforegionInfo,
- ServerNamehostingServer,
- booleanforce,
- booleanremoveAfterUnassigning)
-
-
-UnassignProcedure(RegionInforegionInfo,
- ServerNamehostingServer,
- ServerNamedestinationServer,
- booleanforce)
-
-
-UnassignProcedure(RegionInforegionInfo,
- ServerNamehostingServer,
- ServerNamedestinationServer,
- booleanforce,
- booleanremoveAfterUnassigning)
+UnassignProcedure()
+Deprecated.
+
 
 
 
@@ -274,7 +242,7 @@ extends 
-All MethodsInstance MethodsConcrete Methods
+All 

[23/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.

2018-08-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html
index 8359449..24080ca 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html
@@ -36,104 +36,106 @@
 028import 
org.apache.hadoop.mapreduce.Counter;
 029import 
org.apache.yetus.audience.InterfaceAudience;
 030import 
org.apache.hadoop.conf.Configuration;
-031
-032/**
-033 * Write table content out to map output 
files.
-034 */
-035@InterfaceAudience.Public
-036public class TsvImporterTextMapper
-037extends MapperLongWritable, Text, 
ImmutableBytesWritable, Text
-038{
-039
-040  /** Column seperator */
-041  private String separator;
-042
-043  /** Should skip bad lines */
-044  private boolean skipBadLines;
-045  private Counter badLineCount;
-046  private boolean logBadLines;
-047
-048  private ImportTsv.TsvParser parser;
+031import org.slf4j.Logger;
+032import org.slf4j.LoggerFactory;
+033
+034/**
+035 * Write table content out to map output 
files.
+036 */
+037@InterfaceAudience.Public
+038public class TsvImporterTextMapper
+039extends MapperLongWritable, Text, 
ImmutableBytesWritable, Text {
+040  private static final Logger LOG = 
LoggerFactory.getLogger(TsvImporterTextMapper.class);
+041
+042  /** Column seperator */
+043  private String separator;
+044
+045  /** Should skip bad lines */
+046  private boolean skipBadLines;
+047  private Counter badLineCount;
+048  private boolean logBadLines;
 049
-050  public boolean getSkipBadLines() {
-051return skipBadLines;
-052  }
-053
-054  public Counter getBadLineCount() {
-055return badLineCount;
-056  }
-057
-058  public void incrementBadLineCount(int 
count) {
-059this.badLineCount.increment(count);
-060  }
-061
-062  /**
-063   * Handles initializing this class with 
objects specific to it (i.e., the parser).
-064   * Common initialization that might be 
leveraged by a subclass is done in
-065   * codedoSetup/code. 
Hence a subclass may choose to override this method
-066   * and call 
codedoSetup/code as well before handling it's own custom 
params.
-067   *
-068   * @param context
-069   */
-070  @Override
-071  protected void setup(Context context) 
{
-072doSetup(context);
-073
-074Configuration conf = 
context.getConfiguration();
+050  private ImportTsv.TsvParser parser;
+051
+052  public boolean getSkipBadLines() {
+053return skipBadLines;
+054  }
+055
+056  public Counter getBadLineCount() {
+057return badLineCount;
+058  }
+059
+060  public void incrementBadLineCount(int 
count) {
+061this.badLineCount.increment(count);
+062  }
+063
+064  /**
+065   * Handles initializing this class with 
objects specific to it (i.e., the parser).
+066   * Common initialization that might be 
leveraged by a subclass is done in
+067   * codedoSetup/code. 
Hence a subclass may choose to override this method
+068   * and call 
codedoSetup/code as well before handling it's own custom 
params.
+069   *
+070   * @param context
+071   */
+072  @Override
+073  protected void setup(Context context) 
{
+074doSetup(context);
 075
-076parser = new 
ImportTsv.TsvParser(conf.get(ImportTsv.COLUMNS_CONF_KEY), separator);
-077if (parser.getRowKeyColumnIndex() == 
-1) {
-078  throw new RuntimeException("No row 
key column specified");
-079}
-080  }
-081
-082  /**
-083   * Handles common parameter 
initialization that a subclass might want to leverage.
-084   * @param context
-085   */
-086  protected void doSetup(Context context) 
{
-087Configuration conf = 
context.getConfiguration();
-088
-089// If a custom separator has been 
used,
-090// decode it back from Base64 
encoding.
-091separator = 
conf.get(ImportTsv.SEPARATOR_CONF_KEY);
-092if (separator == null) {
-093  separator = 
ImportTsv.DEFAULT_SEPARATOR;
-094} else {
-095  separator = new 
String(Base64.getDecoder().decode(separator));
-096}
-097
-098skipBadLines = 
context.getConfiguration().getBoolean(ImportTsv.SKIP_LINES_CONF_KEY, true);
-099logBadLines = 
context.getConfiguration().getBoolean(ImportTsv.LOG_BAD_LINES_CONF_KEY, 
false);
-100badLineCount = 
context.getCounter("ImportTsv", "Bad Lines");
-101  }
-102
-103  /**
-104   * Convert a line of TSV text into an 
HBase table row.
-105   */
-106  @Override
-107  public void map(LongWritable offset, 
Text value, Context context) throws IOException {
-108try {
-109  PairInteger,Integer 
rowKeyOffests = parser.parseRowKey(value.getBytes(), value.getLength());
-110  ImmutableBytesWritable rowKey = new 
ImmutableBytesWritable(
-111  

[23/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.

2018-08-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
index b7b4236..3d1edb3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
@@ -259,1863 +259,1867 @@
 251   * + Metadata!  + = See note on 
BLOCK_METADATA_SPACE above.
 252   * ++
 253   * /code
-254   * @see #serialize(ByteBuffer)
+254   * @see #serialize(ByteBuffer, 
boolean)
 255   */
-256  static final 
CacheableDeserializerCacheable BLOCK_DESERIALIZER =
-257  new 
CacheableDeserializerCacheable() {
-258@Override
-259public HFileBlock 
deserialize(ByteBuff buf, boolean reuse, MemoryType memType)
-260throws IOException {
-261  // The buf has the file block 
followed by block metadata.
-262  // Set limit to just before the 
BLOCK_METADATA_SPACE then rewind.
-263  buf.limit(buf.limit() - 
BLOCK_METADATA_SPACE).rewind();
-264  // Get a new buffer to pass the 
HFileBlock for it to 'own'.
-265  ByteBuff newByteBuff;
-266  if (reuse) {
-267newByteBuff = buf.slice();
-268  } else {
-269int len = buf.limit();
-270newByteBuff = new 
SingleByteBuff(ByteBuffer.allocate(len));
-271newByteBuff.put(0, buf, 
buf.position(), len);
-272  }
-273  // Read out the 
BLOCK_METADATA_SPACE content and shove into our HFileBlock.
-274  buf.position(buf.limit());
-275  buf.limit(buf.limit() + 
HFileBlock.BLOCK_METADATA_SPACE);
-276  boolean usesChecksum = buf.get() == 
(byte) 1;
-277  long offset = buf.getLong();
-278  int nextBlockOnDiskSize = 
buf.getInt();
-279  HFileBlock hFileBlock =
-280  new HFileBlock(newByteBuff, 
usesChecksum, memType, offset, nextBlockOnDiskSize, null);
-281  return hFileBlock;
-282}
-283
-284@Override
-285public int 
getDeserialiserIdentifier() {
-286  return DESERIALIZER_IDENTIFIER;
-287}
-288
-289@Override
-290public HFileBlock 
deserialize(ByteBuff b) throws IOException {
-291  // Used only in tests
-292  return deserialize(b, false, 
MemoryType.EXCLUSIVE);
-293}
-294  };
-295
-296  private static final int 
DESERIALIZER_IDENTIFIER;
-297  static {
-298DESERIALIZER_IDENTIFIER =
-299
CacheableDeserializerIdManager.registerDeserializer(BLOCK_DESERIALIZER);
-300  }
-301
-302  /**
-303   * Copy constructor. Creates a shallow 
copy of {@code that}'s buffer.
-304   */
-305  private HFileBlock(HFileBlock that) {
-306this(that, false);
-307  }
-308
-309  /**
-310   * Copy constructor. Creates a 
shallow/deep copy of {@code that}'s buffer as per the boolean
-311   * param.
-312   */
-313  private HFileBlock(HFileBlock that, 
boolean bufCopy) {
-314init(that.blockType, 
that.onDiskSizeWithoutHeader,
-315
that.uncompressedSizeWithoutHeader, that.prevBlockOffset,
-316that.offset, 
that.onDiskDataSizeWithHeader, that.nextBlockOnDiskSize, that.fileContext);
-317if (bufCopy) {
-318  this.buf = new 
SingleByteBuff(ByteBuffer.wrap(that.buf.toBytes(0, that.buf.limit(;
-319} else {
-320  this.buf = that.buf.duplicate();
-321}
-322  }
-323
-324  /**
-325   * Creates a new {@link HFile} block 
from the given fields. This constructor
-326   * is used only while writing blocks 
and caching,
-327   * and is sitting in a byte buffer and 
we want to stuff the block into cache.
-328   *
-329   * pTODO: The caller presumes 
no checksumming
-330   * required of this block instance 
since going into cache; checksum already verified on
-331   * underlying block data pulled in from 
filesystem. Is that correct? What if cache is SSD?
+256  public static final 
CacheableDeserializerCacheable BLOCK_DESERIALIZER = new 
BlockDeserializer();
+257
+258  public static final class 
BlockDeserializer implements CacheableDeserializerCacheable {
+259private BlockDeserializer() {
+260}
+261
+262@Override
+263public HFileBlock 
deserialize(ByteBuff buf, boolean reuse, MemoryType memType)
+264throws IOException {
+265  // The buf has the file block 
followed by block metadata.
+266  // Set limit to just before the 
BLOCK_METADATA_SPACE then rewind.
+267  buf.limit(buf.limit() - 
BLOCK_METADATA_SPACE).rewind();
+268  // Get a new buffer to pass the 
HFileBlock for it to 'own'.
+269  ByteBuff newByteBuff;
+270  if (reuse) {
+271newByteBuff = buf.slice();
+272  } else {
+273int len = buf.limit();
+274newByteBuff = new 
SingleByteBuff(ByteBuffer.allocate(len));
+275newByteBuff.put(0, buf, 
buf.position(), len);

[23/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.

2018-07-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html
index 4bde540..0a56e97 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html
@@ -334,7 +334,7 @@ implements Procedure
-acquireLock,
 addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 d
 oExecute, doReleaseLock,
 doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId, getProcedureMetrics,
 getProcId,
 getProcIdHashCode,
 getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes, getState,
 getSubmittedTime,
 getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent, hasTimeout,
 haveSameParent,
 holdLock,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 i
 sRunnable, isSuccess,
 isWaiting,
 isYieldAfterExecutionStep,
 releaseLock,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure, setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner,
 setOwner,
 setParentProcId,
 setProcId,
 setResult, setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
  toString,
 toStringClass,
 toStringClassDetails,
 toStringDetails,
 toStringSimpleSB,
 toStringState,
 updateMetricsOnFinish,
 updateMetricsOnSubmit, updateTimestamp,
 wasExecuted
+acquireLock,
 addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doExecute,
 doRollba
 ck, elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics,
 getProcId, getProcIdHashCode,
 getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedT
 ime, getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 holdLock,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 isYieldAfterExecutionStep,
 releaseLock,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate, setNonceKey,
 setOwner,
 setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes, setState,
 setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
 toString,
 toStringClass,
 toStringClassDetails,
 toStringDetails,
 toStringSimpleSB,
 toStringState,
 updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 
 waitInitialized, wasExecuted
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html
index b7f19fe..742e018 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html
@@ -299,7 +299,7 @@ extends AbstractPeerProcedure
-acquireLock,
 getLatch,
 getPeerId,
 hasLock,
 holdLock,
 refreshPeer, 
releaseLock,
 rollbackState
+acquireLock,
 getLatch,
 getPeerId,
 holdLock,
 refreshPeer,
 releaseLock,
 rollbackState,
 waitInitialized
 
 
 
@@ -313,7 +313,7 @@ extends Procedure
-addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 doExecute,
 d
 oReleaseLock, doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics, getProcId,
 getProcIdHashCode,
 getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState, getSubmittedTime,
 getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner,
 setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 <
 a 
href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#setTimeout-int-">setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
 toString,
 toStringClass,
 toStringClassDetails,
 toStringDetails,
 toStringSimpleSB,
 updateMetricsOnFinish,
 

[23/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.

2018-07-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html
index 233dba3..91b9055 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html
@@ -540,1205 +540,1204 @@
 532  
sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) - {
 533DequeBalancerRegionLoad 
rLoads = oldLoads.get(Bytes.toString(regionName));
 534if (rLoads == null) {
-535  // There was nothing there
-536  rLoads = new 
ArrayDeque();
-537} else if (rLoads.size() = 
numRegionLoadsToRemember) {
-538  rLoads.remove();
-539}
-540rLoads.add(new 
BalancerRegionLoad(rm));
-541
loads.put(Bytes.toString(regionName), rLoads);
-542  });
-543});
-544
-545for(CostFromRegionLoadFunction cost : 
regionLoadFunctions) {
-546  cost.setLoads(loads);
-547}
-548  }
-549
-550  protected void initCosts(Cluster 
cluster) {
-551for (CostFunction c:costFunctions) 
{
-552  c.init(cluster);
-553}
-554  }
-555
-556  protected void 
updateCostsWithAction(Cluster cluster, Action action) {
-557for (CostFunction c : costFunctions) 
{
-558  c.postAction(action);
-559}
-560  }
-561
-562  /**
-563   * Get the names of the cost 
functions
-564   */
-565  public String[] getCostFunctionNames() 
{
-566if (costFunctions == null) return 
null;
-567String[] ret = new 
String[costFunctions.length];
-568for (int i = 0; i  
costFunctions.length; i++) {
-569  CostFunction c = 
costFunctions[i];
-570  ret[i] = 
c.getClass().getSimpleName();
-571}
-572
-573return ret;
-574  }
-575
-576  /**
-577   * This is the main cost function.  It 
will compute a cost associated with a proposed cluster
-578   * state.  All different costs will be 
combined with their multipliers to produce a double cost.
-579   *
-580   * @param cluster The state of the 
cluster
-581   * @param previousCost the previous 
cost. This is used as an early out.
-582   * @return a double of a cost 
associated with the proposed cluster state.  This cost is an
-583   * aggregate of all individual 
cost functions.
-584   */
-585  protected double computeCost(Cluster 
cluster, double previousCost) {
-586double total = 0;
-587
-588for (int i = 0; i  
costFunctions.length; i++) {
-589  CostFunction c = 
costFunctions[i];
-590  this.tempFunctionCosts[i] = 0.0;
-591
-592  if (c.getMultiplier() = 0) {
-593continue;
-594  }
-595
-596  Float multiplier = 
c.getMultiplier();
-597  Double cost = c.cost();
-598
-599  this.tempFunctionCosts[i] = 
multiplier*cost;
-600  total += 
this.tempFunctionCosts[i];
-601
-602  if (total  previousCost) {
-603break;
-604  }
-605}
-606
-607return total;
-608  }
-609
-610  /** Generates a candidate action to be 
applied to the cluster for cost function search */
-611  abstract static class 
CandidateGenerator {
-612abstract Cluster.Action 
generate(Cluster cluster);
-613
-614/**
-615 * From a list of regions pick a 
random one. Null can be returned which
-616 * {@link 
StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region 
move
-617 * rather than swap.
-618 *
-619 * @param clusterThe state of 
the cluster
-620 * @param server index of the 
server
-621 * @param chanceOfNoSwap Chance that 
this will decide to try a move rather
-622 *   than a 
swap.
-623 * @return a random {@link 
RegionInfo} or null if an asymmetrical move is
-624 * suggested.
-625 */
-626protected int 
pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) {
-627  // Check to see if this is just a 
move.
-628  if 
(cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat()  
chanceOfNoSwap) {
-629// signal a move only.
-630return -1;
-631  }
-632  int rand = 
RANDOM.nextInt(cluster.regionsPerServer[server].length);
-633  return 
cluster.regionsPerServer[server][rand];
-634
-635}
-636protected int 
pickRandomServer(Cluster cluster) {
-637  if (cluster.numServers  1) {
-638return -1;
-639  }
-640
-641  return 
RANDOM.nextInt(cluster.numServers);
-642}
-643
-644protected int pickRandomRack(Cluster 
cluster) {
-645  if (cluster.numRacks 

[23/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.

2018-07-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html 
b/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html
index 1665f28..fd82164 100644
--- a/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html
+++ b/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html
@@ -1,6 +1,6 @@
 http://www.w3.org/TR/html4/loose.dtd;>
 
-
+
 
 
 
@@ -20,38 +20,38 @@
 //-->
 
 
-您的浏览器已禁用 JavaScript。
+JavaScript is disabled on your browser.
 
 
 
 
 
-跳过导航链接
+Skip navigation links
 
 
 
-
-概览
-程序包
-ç±»
-使用
-树
-已过时
-索引
-帮助
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
 
 
 
 
-上一个类
-下一个类
+PrevClass
+NextClass
 
 
-框架
-无框架
+Frames
+NoFrames
 
 
-所有类
+AllClasses
 
 
 
 
org.apache.hadoop.hbase
-

ç±» NotServingRegionException

+

Class NotServingRegionException


[23/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.

2018-07-19 Thread zhangduo
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html 
b/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html
index e655215..ba2dc6f 100644
--- a/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html
+++ b/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html
@@ -1,6 +1,6 @@
 http://www.w3.org/TR/html4/loose.dtd;>
 
-
+
 
 
 
@@ -20,38 +20,38 @@
 //-->
 
 
-JavaScript is disabled on your browser.
+您的浏览器已禁用 JavaScript。
 
 
 
 
 
-Skip navigation links
+跳过导航链接
 
 
 
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
+
+概览
+程序包
+ç±»
+使用
+树
+已过时
+索引
+帮助
 
 
 
 
-PrevClass
-NextClass
+上一个类
+下一个类
 
 
-Frames
-NoFrames
+框架
+无框架
 
 
-AllClasses
+所有类
 
 
 
 
org.apache.hadoop.hbase
-

Class NamespaceExistException

+

ç±» NamespaceExistException


[23/51] [partial] hbase-site git commit: Published site at 0f23784182ab88649de340d75804e0ff20dcd0fc.

2018-07-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html
 
b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html
index 432cba4..8851d0f 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class ReplicationSource
+public class ReplicationSource
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements ReplicationSourceInterface
 Class that handles the source of a replication stream.
@@ -369,55 +369,65 @@ implements 
+private long
+getFileSize(org.apache.hadoop.fs.PathcurrentPath)
+
+
 ReplicationPeer
 getPeer()
 Get the replication peer instance.
 
 
-
+
 https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 getQueueId()
 Get the queue id that the source is replicating to
 
 
-
+
 (package private) ReplicationQueueStorage
 getQueueStorage()
 
-
+
 ReplicationEndpoint
 getReplicationEndpoint()
 
-
+
 (package private) Server
 getServer()
 
-
+
 ServerName
 getServerWALsBelongTo()
 The queue of WALs only belong to one region server.
 
 
-
+
 ReplicationSourceManager
 getSourceManager()
 
-
+
 MetricsSource
 getSourceMetrics()
 
-
+
 https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 getStats()
 Get a string representation of the current statistics
  for this source
 
 
-
+
 WALFileLengthProvider
 getWALFileLengthProvider()
 
-
+
+https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,ReplicationStatus
+getWalGroupStatus()
+get the stat of replication for each wal group.
+
+
+
 void
 init(org.apache.hadoop.conf.Configurationconf,
 org.apache.hadoop.fs.FileSystemfs,
@@ -432,77 +442,77 @@ implements Instantiation method used by region servers
 
 
-
+
 private void
 initAndStartReplicationEndpoint(ReplicationEndpointreplicationEndpoint)
 
-
+
 private void
 initialize()
 
-
+
 private void
 initializeWALEntryFilter(https://docs.oracle.com/javase/8/docs/api/java/util/UUID.html?is-external=true;
 title="class or interface in 
java.util">UUIDpeerClusterId)
 
-
+
 boolean
 isSourceActive()
 
-
+
 void
 postShipEdits(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListWAL.Entryentries,
  intbatchSize)
 Call this after the shipper thread ship some entries to 
peer cluster.
 
 
-
+
 (package private) void
 removeWorker(ReplicationSourceShipperworker)
 
-
+
 protected boolean
 sleepForRetries(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringmsg,
intsleepMultiplier)
 Do the sleeping logic
 
 
-
+
 void
 startup()
 Start the replication
 
 
-
+
 void
 terminate(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringreason)
 End the replication
 
 
-
+
 void
 terminate(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringreason,
  https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exceptioncause)
 End the replication
 
 
-
+
 void
 terminate(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 

[23/51] [partial] hbase-site git commit: Published site at 85b41f36e01214b6485c9352875c84ebf877dab3.

2018-06-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html
 
b/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html
index a1f615f..9ca96d9 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html
@@ -506,10 +506,47 @@ the order they are declared.
 
 
 
+static boolean
+AccessControlClient.hasPermission(Connectionconnection,
+ https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtableName,
+ byte[]columnFamily,
+ byte[]columnQualifier,
+ https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringuserName,
+ Permission.Action...actions)
+Validates whether specified user has permission to perform 
actions on the mentioned table,
+ column family or column qualifier.
+
+
+
+static boolean
+AccessControlClient.hasPermission(Connectionconnection,
+ https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtableName,
+ https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringcolumnFamily,
+ https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringcolumnQualifier,
+ https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringuserName,
+ Permission.Action...actions)
+Validates whether specified user has permission to perform 
actions on the mentioned table,
+ column family or column qualifier.
+
+
+
+static boolean
+AccessControlUtil.hasPermission(com.google.protobuf.RpcControllercontroller,
+ 
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol,
+ TableNametableName,
+ byte[]columnFamily,
+ byte[]columnQualifier,
+ https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringuserName,
+ Permission.Action[]actions)
+Validates whether specified user has permission to perform 
actions on the mentioned table,
+ column family or column qualifier.
+
+
+
 boolean
 Permission.implies(Permission.Actionaction)
 
-
+
 boolean
 TablePermission.implies(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringnamespace,
Permission.Actionaction)
@@ -517,7 +554,7 @@ the order they are declared.
  instance.
 
 
-
+
 boolean
 TablePermission.implies(TableNametable,
byte[]family,
@@ -527,7 +564,7 @@ the order they are declared.
  instance.
 
 
-
+
 boolean
 TablePermission.implies(TableNametable,
KeyValuekv,
@@ -536,7 +573,7 @@ the order they are declared.
  the given table and key value.
 
 
-
+
 boolean
 TablePermission.matchesFamily(TableNametable,
  byte[]family,
@@ -545,7 +582,7 @@ the order they are declared.
  family at least.
 
 
-
+
 boolean
 TablePermission.matchesFamilyQualifier(TableNametable,
   byte[]family,
@@ -554,7 +591,7 @@ the order they are declared.
 Returns if the given permission matches the given 
qualifier.
 
 
-
+
 boolean
 TableAuthManager.matchPermission(Useruser,
TableNametable,
@@ -562,7 +599,7 @@ the order they are declared.
byte[]qualifier,
Permission.Actionaction)
 
-
+
 boolean
 TableAuthManager.matchPermission(Useruser,
TableNametable,
@@ -572,7 +609,7 @@ the order they are declared.
  to the column family portion of a permission.
 
 
-
+
 private AuthResult
 AccessController.permissionGranted(AccessController.OpTypeopType,
  Useruser,
@@ -583,25 +620,26 @@ the order they are declared.
  against the given set of row data.
 
 
-
+
 private AuthResult
-AccessController.permissionGranted(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringrequest,
+AccessController.permissionGranted(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringrequest,
  Useruser,
  Permission.ActionpermRequest,
  RegionCoprocessorEnvironmente,
+ 

[23/51] [partial] hbase-site git commit: Published site at 6198e1fc7dfa85c3bc6b2855f9a5fb5f4b2354ff.

2018-06-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
index 37f4bd5..c84d49c 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i
 
109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":10,"i175":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i
 
109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":10,"i175":10,"i176":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-class RawAsyncHBaseAdmin
+class RawAsyncHBaseAdmin
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements AsyncAdmin
 The implementation of AsyncAdmin.
@@ -1355,49 +1355,56 @@ implements 
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
+transitReplicationPeerSyncReplicationState(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringpeerId,
+  

[23/51] [partial] hbase-site git commit: Published site at 14087cc919da9f2e0b1a68f701f6365ad9d1d71f.

2018-06-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html
index ccbd322..d044d2e 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10};
+var methods = 
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -191,10 +191,6 @@ extends modifiedTableDescriptor
 
 
-private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
-traceEnabled
-
-
 private TableDescriptor
 unmodifiedTableDescriptor
 
@@ -288,103 +284,85 @@ extends 
 private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-getOpenRegionInfoList(MasterProcedureEnvenv)
-Fetches all open or soon to be open Regions for a 
table.
-
-
-
-private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
 getRegionInfoList(MasterProcedureEnvenv)
 Fetches all Regions for a table.
 
 
-
+
 protected 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableState
 getState(intstateId)
 Convert an ordinal (or state id) to an Enum (or more 
descriptive) state object.
 
 
-
+
 protected int
 getStateId(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)
 Convert the Enum (or more descriptive) state object to an 
ordinal (or state id).
 
 
-
+
 TableName
 getTableName()
 
-
+
 TableProcedureInterface.TableOperationType
 getTableOperationType()
 Given an operation type we can take decisions about what to 
do with pending operations.
 
 
-
+
 private void
 initilize()
 
-
+
 protected boolean
 isRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)
 Used by the default implementation of abort() to know if 
the current state can be aborted
  and rollback can be triggered.
 
 
-
-private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
-isTraceEnabled()
-The procedure could be restarted from a different 
machine.
-
-
-
+
 private void
 postModify(MasterProcedureEnvenv,
   
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)
 Action after modifying table.
 
 
-
+
 private void
 preModify(MasterProcedureEnvenv,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)
 Action before modifying table.
 
 
-
+
 private void
 prepareModify(MasterProcedureEnvenv)
 Check conditions before any real action of modifying a 
table.
 
 
-
-private void
-restoreTableDescriptor(MasterProcedureEnvenv)
-Undo the descriptor change (for rollback)
-
-
-
+
 protected void
 rollbackState(MasterProcedureEnvenv,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)
 called to perform the rollback of the specified state
 
 
-
+
 private void
 runCoprocessorAction(MasterProcedureEnvenv,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)
 Coprocessor Action.
 
 
-
+
 protected void
 serializeStateData(ProcedureStateSerializerserializer)
 The user-level code of the procedure may have some state to
  persist (e.g.
 
 
-
+
 private void
 updateReplicaColumnsIfNeeded(MasterProcedureEnvenv,
 TableDescriptoroldTableDescriptor,
@@ -392,7 +370,7 @@ extends update replica column families if necessary.
 
 
-
+
 private void
 updateTableDescriptor(MasterProcedureEnvenv)
 Update descriptor
@@ -471,21 +449,12 @@ extends 
 
 
-
+
 
 deleteColumnFamilyInModify
 privateboolean deleteColumnFamilyInModify
 
 
-
-
-
-
-
-traceEnabled
-privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
-
-
 
 
 
@@ -500,7 

[23/51] [partial] hbase-site git commit: Published site at 72784c2d836a4b977667449d3adec5e8d15453f5.

2018-06-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
index b6e7636..592c2cc 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
@@ -356,3901 +356,3924 @@
 348  public FutureVoid 
modifyTableAsync(TableDescriptor td) throws IOException {
 349ModifyTableResponse response = 
executeCallable(
 350  new 
MasterCallableModifyTableResponse(getConnection(), 
getRpcControllerFactory()) {
-351@Override
-352protected ModifyTableResponse 
rpcCall() throws Exception {
-353  
setPriority(td.getTableName());
-354  ModifyTableRequest request = 
RequestConverter.buildModifyTableRequest(
-355td.getTableName(), td, 
ng.getNonceGroup(), ng.newNonce());
-356  return 
master.modifyTable(getRpcController(), request);
-357}
-358  });
-359return new ModifyTableFuture(this, 
td.getTableName(), response);
-360  }
-361
-362  @Override
-363  public ListTableDescriptor 
listTableDescriptorsByNamespace(byte[] name) throws IOException {
-364return executeCallable(new 
MasterCallableListTableDescriptor(getConnection(),
-365getRpcControllerFactory()) {
-366  @Override
-367  protected 
ListTableDescriptor rpcCall() throws Exception {
-368return 
master.listTableDescriptorsByNamespace(getRpcController(),
-369
ListTableDescriptorsByNamespaceRequest.newBuilder()
-370  
.setNamespaceName(Bytes.toString(name)).build())
-371.getTableSchemaList()
-372.stream()
-373
.map(ProtobufUtil::toTableDescriptor)
-374
.collect(Collectors.toList());
-375  }
-376});
-377  }
-378
-379  @Override
-380  public ListTableDescriptor 
listTableDescriptors(ListTableName tableNames) throws IOException {
-381return executeCallable(new 
MasterCallableListTableDescriptor(getConnection(),
-382getRpcControllerFactory()) {
-383  @Override
-384  protected 
ListTableDescriptor rpcCall() throws Exception {
-385GetTableDescriptorsRequest req 
=
-386
RequestConverter.buildGetTableDescriptorsRequest(tableNames);
-387  return 
ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(),
-388  req));
-389  }
-390});
-391  }
-392
-393  @Override
-394  public ListRegionInfo 
getRegions(final ServerName sn) throws IOException {
-395AdminService.BlockingInterface admin 
= this.connection.getAdmin(sn);
-396// TODO: There is no timeout on this 
controller. Set one!
-397HBaseRpcController controller = 
rpcControllerFactory.newController();
-398return 
ProtobufUtil.getOnlineRegions(controller, admin);
-399  }
-400
-401  @Override
-402  public ListRegionInfo 
getRegions(TableName tableName) throws IOException {
-403if 
(TableName.isMetaTableName(tableName)) {
-404  return 
Arrays.asList(RegionInfoBuilder.FIRST_META_REGIONINFO);
-405} else {
-406  return 
MetaTableAccessor.getTableRegions(connection, tableName, true);
-407}
-408  }
-409
-410  private static class 
AbortProcedureFuture extends ProcedureFutureBoolean {
-411private boolean isAbortInProgress;
-412
-413public AbortProcedureFuture(
-414final HBaseAdmin admin,
-415final Long procId,
-416final Boolean abortProcResponse) 
{
-417  super(admin, procId);
-418  this.isAbortInProgress = 
abortProcResponse;
-419}
-420
-421@Override
-422public Boolean get(long timeout, 
TimeUnit unit)
-423throws InterruptedException, 
ExecutionException, TimeoutException {
-424  if (!this.isAbortInProgress) {
-425return false;
-426  }
-427  super.get(timeout, unit);
-428  return true;
-429}
-430  }
-431
-432  /** @return Connection used by this 
object. */
-433  @Override
-434  public Connection getConnection() {
-435return connection;
-436  }
-437
-438  @Override
-439  public boolean tableExists(final 
TableName tableName) throws IOException {
-440return executeCallable(new 
RpcRetryingCallableBoolean() {
-441  @Override
-442  protected Boolean rpcCall(int 
callTimeout) throws Exception {
-443return 
MetaTableAccessor.tableExists(connection, tableName);
-444  }
-445});
-446  }
-447
-448  @Override
-449  public HTableDescriptor[] listTables() 
throws IOException {
-450return listTables((Pattern)null, 
false);
-451  }
-452
-453  @Override
-454  public HTableDescriptor[] 
listTables(Pattern pattern) throws IOException {
-455   

[23/51] [partial] hbase-site git commit: Published site at 9101fc246f86445006bfbcdfda5cc495016dc280.

2018-06-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
index fea2b5a..c7a6cc4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
@@ -1354,816 +1354,824 @@
 1346   */
 1347  public static void 
putsToMetaTable(final Connection connection, final ListPut ps)
 1348  throws IOException {
-1349try (Table t = 
getMetaHTable(connection)) {
-1350  debugLogMutations(ps);
-1351  t.put(ps);
-1352}
-1353  }
-1354
-1355  /**
-1356   * Delete the passed 
coded/code from the codehbase:meta/code 
table.
-1357   * @param connection connection we're 
using
-1358   * @param d Delete to add to 
hbase:meta
-1359   */
-1360  private static void 
deleteFromMetaTable(final Connection connection, final Delete d)
-1361  throws IOException {
-1362ListDelete dels = new 
ArrayList(1);
-1363dels.add(d);
-1364deleteFromMetaTable(connection, 
dels);
-1365  }
-1366
-1367  /**
-1368   * Delete the passed 
codedeletes/code from the codehbase:meta/code 
table.
-1369   * @param connection connection we're 
using
-1370   * @param deletes Deletes to add to 
hbase:meta  This list should support #remove.
-1371   */
-1372  private static void 
deleteFromMetaTable(final Connection connection, final ListDelete 
deletes)
-1373  throws IOException {
-1374try (Table t = 
getMetaHTable(connection)) {
-1375  debugLogMutations(deletes);
-1376  t.delete(deletes);
-1377}
-1378  }
-1379
-1380  /**
-1381   * Deletes some replica columns 
corresponding to replicas for the passed rows
-1382   * @param metaRows rows in 
hbase:meta
-1383   * @param replicaIndexToDeleteFrom the 
replica ID we would start deleting from
-1384   * @param numReplicasToRemove how many 
replicas to remove
-1385   * @param connection connection we're 
using to access meta table
-1386   */
-1387  public static void 
removeRegionReplicasFromMeta(Setbyte[] metaRows,
-1388int replicaIndexToDeleteFrom, int 
numReplicasToRemove, Connection connection)
-1389  throws IOException {
-1390int absoluteIndex = 
replicaIndexToDeleteFrom + numReplicasToRemove;
-1391for (byte[] row : metaRows) {
-1392  long now = 
EnvironmentEdgeManager.currentTime();
-1393  Delete deleteReplicaLocations = 
new Delete(row);
-1394  for (int i = 
replicaIndexToDeleteFrom; i  absoluteIndex; i++) {
-1395
deleteReplicaLocations.addColumns(getCatalogFamily(),
-1396  getServerColumn(i), now);
-1397
deleteReplicaLocations.addColumns(getCatalogFamily(),
-1398  getSeqNumColumn(i), now);
-1399
deleteReplicaLocations.addColumns(getCatalogFamily(),
-1400  getStartCodeColumn(i), now);
-1401  }
-1402  deleteFromMetaTable(connection, 
deleteReplicaLocations);
-1403}
-1404  }
-1405
-1406  /**
-1407   * Execute the passed 
codemutations/code against codehbase:meta/code 
table.
-1408   * @param connection connection we're 
using
-1409   * @param mutations Puts and Deletes 
to execute on hbase:meta
-1410   * @throws IOException
-1411   */
-1412  public static void 
mutateMetaTable(final Connection connection,
-1413 
final ListMutation mutations)
-1414throws IOException {
-1415Table t = 
getMetaHTable(connection);
-1416try {
-1417  debugLogMutations(mutations);
-1418  t.batch(mutations, null);
-1419} catch (InterruptedException e) {
-1420  InterruptedIOException ie = new 
InterruptedIOException(e.getMessage());
-1421  ie.initCause(e);
-1422  throw ie;
-1423} finally {
-1424  t.close();
-1425}
-1426  }
-1427
-1428  private static void 
addRegionStateToPut(Put put, RegionState.State state) throws IOException {
-1429
put.add(CellBuilderFactory.create(CellBuilderType.SHALLOW_COPY)
-1430.setRow(put.getRow())
-1431
.setFamily(HConstants.CATALOG_FAMILY)
-1432
.setQualifier(getRegionStateColumn())
-1433
.setTimestamp(put.getTimestamp())
-1434.setType(Cell.Type.Put)
-1435
.setValue(Bytes.toBytes(state.name()))
-1436.build());
-1437  }
-1438
-1439  /**
-1440   * Adds daughter region infos to 
hbase:meta row for the specified region. Note that this does not
-1441   * add its daughter's as different 
rows, but adds information about the daughters in the same row
-1442   * as the parent. Use
-1443   * {@link #splitRegion(Connection, 
RegionInfo, long, RegionInfo, RegionInfo, ServerName, int)}
-1444   * if you want to do that.
-1445   * @param connection connection we're 
using
-1446   * 

[23/51] [partial] hbase-site git commit: Published site at 0b28155d274910b4e667b949d51f78809a1eff0b.

2018-06-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
index 42d0637..eb16038 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html
@@ -80,21 +80,21 @@
 072import 
org.apache.hadoop.hbase.PleaseHoldException;
 073import 
org.apache.hadoop.hbase.ReplicationPeerNotFoundException;
 074import 
org.apache.hadoop.hbase.ScheduledChore;
-075import 
org.apache.hadoop.hbase.ServerMetricsBuilder;
-076import 
org.apache.hadoop.hbase.ServerName;
-077import 
org.apache.hadoop.hbase.TableDescriptors;
-078import 
org.apache.hadoop.hbase.TableName;
-079import 
org.apache.hadoop.hbase.TableNotDisabledException;
-080import 
org.apache.hadoop.hbase.TableNotFoundException;
-081import 
org.apache.hadoop.hbase.UnknownRegionException;
-082import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
-083import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
-084import 
org.apache.hadoop.hbase.client.MasterSwitchType;
-085import 
org.apache.hadoop.hbase.client.RegionInfo;
-086import 
org.apache.hadoop.hbase.client.Result;
-087import 
org.apache.hadoop.hbase.client.TableDescriptor;
-088import 
org.apache.hadoop.hbase.client.TableDescriptorBuilder;
-089import 
org.apache.hadoop.hbase.client.TableState;
+075import 
org.apache.hadoop.hbase.ServerName;
+076import 
org.apache.hadoop.hbase.TableDescriptors;
+077import 
org.apache.hadoop.hbase.TableName;
+078import 
org.apache.hadoop.hbase.TableNotDisabledException;
+079import 
org.apache.hadoop.hbase.TableNotFoundException;
+080import 
org.apache.hadoop.hbase.UnknownRegionException;
+081import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
+082import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
+083import 
org.apache.hadoop.hbase.client.MasterSwitchType;
+084import 
org.apache.hadoop.hbase.client.RegionInfo;
+085import 
org.apache.hadoop.hbase.client.Result;
+086import 
org.apache.hadoop.hbase.client.TableDescriptor;
+087import 
org.apache.hadoop.hbase.client.TableDescriptorBuilder;
+088import 
org.apache.hadoop.hbase.client.TableState;
+089import 
org.apache.hadoop.hbase.client.VersionInfoUtil;
 090import 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
 091import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
 092import 
org.apache.hadoop.hbase.exceptions.MergeRegionException;
@@ -220,3477 +220,3481 @@
 212
 213import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
 214import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState;
-215import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionServerInfo;
-216import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas;
-217import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceViolationPolicy;
-218import 
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;
-219import 
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
-220
-221/**
-222 * HMaster is the "master server" for 
HBase. An HBase cluster has one active
-223 * master.  If many masters are started, 
all compete.  Whichever wins goes on to
-224 * run the cluster.  All others park 
themselves in their constructor until
-225 * master or cluster shutdown or until 
the active master loses its lease in
-226 * zookeeper.  Thereafter, all running 
master jostle to take over master role.
-227 *
-228 * pThe Master can be asked 
shutdown the cluster. See {@link #shutdown()}.  In
-229 * this case it will tell all 
regionservers to go down and then wait on them
-230 * all reporting in that they are down.  
This master will then shut itself down.
-231 *
-232 * pYou can also shutdown just 
this master.  Call {@link #stopMaster()}.
-233 *
-234 * @see org.apache.zookeeper.Watcher
-235 */
-236@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
-237@SuppressWarnings("deprecation")
-238public class HMaster extends 
HRegionServer implements MasterServices {
-239  private static Logger LOG = 
LoggerFactory.getLogger(HMaster.class.getName());
-240
-241  /**
-242   * Protection against zombie master. 
Started once Master accepts active responsibility and
-243   * starts taking over responsibilities. 
Allows a finite time window before giving up ownership.
-244   */
-245  private static class 
InitializationMonitor extends HasThread {
-246/** The amount of time in 
milliseconds to sleep before checking initialization status. */
-247public static final String 
TIMEOUT_KEY = 

[23/51] [partial] hbase-site git commit: Published site at 7d3750bd9fc9747623549c242cc4171e224b3eaf.

2018-06-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
index fe1e077..90c31f4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
@@ -1072,894 +1072,913 @@
 1064
 1065  protected boolean 
waitServerReportEvent(final ServerName serverName, final Procedure proc) {
 1066final ServerStateNode serverNode = 
regionStates.getOrCreateServer(serverName);
-1067return 
serverNode.getReportEvent().suspendIfNotReady(proc);
-1068  }
-1069
-1070  protected void 
wakeServerReportEvent(final ServerStateNode serverNode) {
-1071
serverNode.getReportEvent().wake(getProcedureScheduler());
-1072  }
-1073
-1074  // 

-1075  //  RIT chore
-1076  // 

-1077  private static class 
RegionInTransitionChore extends 
ProcedureInMemoryChoreMasterProcedureEnv {
-1078public RegionInTransitionChore(final 
int timeoutMsec) {
-1079  super(timeoutMsec);
-1080}
-1081
-1082@Override
-1083protected void periodicExecute(final 
MasterProcedureEnv env) {
-1084  final AssignmentManager am = 
env.getAssignmentManager();
-1085
-1086  final RegionInTransitionStat 
ritStat = am.computeRegionInTransitionStat();
-1087  if 
(ritStat.hasRegionsOverThreshold()) {
-1088for (RegionState hri: 
ritStat.getRegionOverThreshold()) {
-1089  
am.handleRegionOverStuckWarningThreshold(hri.getRegion());
-1090}
-1091  }
-1092
-1093  // update metrics
-1094  
am.updateRegionsInTransitionMetrics(ritStat);
-1095}
-1096  }
-1097
-1098  public RegionInTransitionStat 
computeRegionInTransitionStat() {
-1099final RegionInTransitionStat rit = 
new RegionInTransitionStat(getConfiguration());
-1100rit.update(this);
-1101return rit;
-1102  }
-1103
-1104  public static class 
RegionInTransitionStat {
-1105private final int ritThreshold;
+1067if (serverNode == null) {
+1068  LOG.warn("serverName=null; {}", 
proc);
+1069}
+1070return 
serverNode.getReportEvent().suspendIfNotReady(proc);
+1071  }
+1072
+1073  protected void 
wakeServerReportEvent(final ServerStateNode serverNode) {
+1074
serverNode.getReportEvent().wake(getProcedureScheduler());
+1075  }
+1076
+1077  // 

+1078  //  RIT chore
+1079  // 

+1080  private static class 
RegionInTransitionChore extends 
ProcedureInMemoryChoreMasterProcedureEnv {
+1081public RegionInTransitionChore(final 
int timeoutMsec) {
+1082  super(timeoutMsec);
+1083}
+1084
+1085@Override
+1086protected void periodicExecute(final 
MasterProcedureEnv env) {
+1087  final AssignmentManager am = 
env.getAssignmentManager();
+1088
+1089  final RegionInTransitionStat 
ritStat = am.computeRegionInTransitionStat();
+1090  if 
(ritStat.hasRegionsOverThreshold()) {
+1091for (RegionState hri: 
ritStat.getRegionOverThreshold()) {
+1092  
am.handleRegionOverStuckWarningThreshold(hri.getRegion());
+1093}
+1094  }
+1095
+1096  // update metrics
+1097  
am.updateRegionsInTransitionMetrics(ritStat);
+1098}
+1099  }
+1100
+1101  public RegionInTransitionStat 
computeRegionInTransitionStat() {
+1102final RegionInTransitionStat rit = 
new RegionInTransitionStat(getConfiguration());
+1103rit.update(this);
+1104return rit;
+1105  }
 1106
-1107private HashMapString, 
RegionState ritsOverThreshold = null;
-1108private long statTimestamp;
-1109private long oldestRITTime = 0;
-1110private int totalRITsTwiceThreshold 
= 0;
-private int totalRITs = 0;
-1112
-1113@VisibleForTesting
-1114public RegionInTransitionStat(final 
Configuration conf) {
-1115  this.ritThreshold =
-1116
conf.getInt(METRICS_RIT_STUCK_WARNING_THRESHOLD, 
DEFAULT_RIT_STUCK_WARNING_THRESHOLD);
-1117}
-1118
-1119public int getRITThreshold() {
-1120  return ritThreshold;
-1121}
-1122
-1123public long getTimestamp() {
-1124  return statTimestamp;
-1125}
-1126
-1127public int getTotalRITs() {
-1128  return totalRITs;
-1129}
-1130
-1131public long getOldestRITTime() {

[23/51] [partial] hbase-site git commit: Published site at 997747076d8ec0b4346d7cb99c4b0667a7c14905.

2018-05-30 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
index 3da432b..d30fa8f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
@@ -928,7690 +928,7698 @@
 920  CollectionHStore stores = 
this.stores.values();
 921  try {
 922// update the stores that we are 
replaying
-923
stores.forEach(HStore::startReplayingFromWAL);
-924// Recover any edits if 
available.
-925maxSeqId = Math.max(maxSeqId,
-926  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-927// Make sure mvcc is up to max.
-928this.mvcc.advanceTo(maxSeqId);
-929  } finally {
-930// update the stores that we are 
done replaying
-931
stores.forEach(HStore::stopReplayingFromWAL);
-932  }
-933}
-934this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-935
-936
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-937this.writestate.flushRequested = 
false;
-938this.writestate.compacting.set(0);
-939
-940if (this.writestate.writesEnabled) 
{
-941  // Remove temporary data left over 
from old regions
-942  status.setStatus("Cleaning up 
temporary data from old regions");
-943  fs.cleanupTempDir();
-944}
-945
-946if (this.writestate.writesEnabled) 
{
-947  status.setStatus("Cleaning up 
detritus from prior splits");
-948  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-949  // these directories here on open.  
We may be opening a region that was
-950  // being split but we crashed in 
the middle of it all.
-951  fs.cleanupAnySplitDetritus();
-952  fs.cleanupMergesDir();
-953}
-954
-955// Initialize split policy
-956this.splitPolicy = 
RegionSplitPolicy.create(this, conf);
-957
-958// Initialize flush policy
-959this.flushPolicy = 
FlushPolicyFactory.create(this, conf);
-960
-961long lastFlushTime = 
EnvironmentEdgeManager.currentTime();
-962for (HStore store: stores.values()) 
{
-963  
this.lastStoreFlushTimeMap.put(store, lastFlushTime);
-964}
-965
-966// Use maximum of log sequenceid or 
that which was found in stores
-967// (particularly if no recovered 
edits, seqid will be -1).
-968long maxSeqIdFromFile =
-969  
WALSplitter.getMaxRegionSequenceId(fs.getFileSystem(), fs.getRegionDir());
-970long nextSeqId = Math.max(maxSeqId, 
maxSeqIdFromFile) + 1;
-971if (writestate.writesEnabled) {
-972  
WALSplitter.writeRegionSequenceIdFile(fs.getFileSystem(), fs.getRegionDir(), 
nextSeqId - 1);
-973}
-974
-975LOG.info("Opened {}; next 
sequenceid={}", this.getRegionInfo().getShortNameToLog(), nextSeqId);
-976
-977// A region can be reopened if failed 
a split; reset flags
-978this.closing.set(false);
-979this.closed.set(false);
-980
-981if (coprocessorHost != null) {
-982  status.setStatus("Running 
coprocessor post-open hooks");
-983  coprocessorHost.postOpen();
-984}
+923LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
+924
stores.forEach(HStore::startReplayingFromWAL);
+925// Recover any edits if 
available.
+926maxSeqId = Math.max(maxSeqId,
+927  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
+928// Make sure mvcc is up to max.
+929this.mvcc.advanceTo(maxSeqId);
+930  } finally {
+931LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
+932// update the stores that we are 
done replaying
+933
stores.forEach(HStore::stopReplayingFromWAL);
+934  }
+935}
+936this.lastReplayedOpenRegionSeqId = 
maxSeqId;
+937
+938
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
+939this.writestate.flushRequested = 
false;
+940this.writestate.compacting.set(0);
+941
+942if (this.writestate.writesEnabled) 
{
+943  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
+944  // Remove temporary data left over 
from old regions
+945  status.setStatus("Cleaning up 
temporary data from old regions");
+946  fs.cleanupTempDir();
+947}
+948
+949if (this.writestate.writesEnabled) 
{
+950  status.setStatus("Cleaning up 
detritus from prior splits");
+951  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
+952  // these 

[23/51] [partial] hbase-site git commit: Published site at f3d1c021de2264301f68eadb9ef126ff83d7ef53.

2018-05-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html
index 9644187..b979909 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html
@@ -1199,2602 +1199,2599 @@
 1191ClusterStatusProtos.ServerLoad sl = 
buildServerLoad(reportStartTime, reportEndTime);
 1192try {
 1193  RegionServerReportRequest.Builder 
request = RegionServerReportRequest.newBuilder();
-1194  ServerName sn = 
ServerName.parseVersionedServerName(this.serverName.getVersionedBytes());
-1195  
request.setServer(ProtobufUtil.toServerName(sn));
-1196  request.setLoad(sl);
-1197  rss.regionServerReport(null, 
request.build());
-1198} catch (ServiceException se) {
-1199  IOException ioe = 
ProtobufUtil.getRemoteException(se);
-1200  if (ioe instanceof 
YouAreDeadException) {
-1201// This will be caught and 
handled as a fatal error in run()
-1202throw ioe;
-1203  }
-1204  if (rssStub == rss) {
-1205rssStub = null;
-1206  }
-1207  // Couldn't connect to the master, 
get location from zk and reconnect
-1208  // Method blocks until new master 
is found or we are stopped
-1209  
createRegionServerStatusStub(true);
-1210}
-1211  }
-1212
-1213  /**
-1214   * Reports the given map of Regions 
and their size on the filesystem to the active Master.
-1215   *
-1216   * @param regionSizeStore The store 
containing region sizes
-1217   * @return false if 
FileSystemUtilizationChore should pause reporting to master. true otherwise
-1218   */
-1219  public boolean 
reportRegionSizesForQuotas(RegionSizeStore regionSizeStore) {
-1220
RegionServerStatusService.BlockingInterface rss = rssStub;
-1221if (rss == null) {
-1222  // the current server could be 
stopping.
-1223  LOG.trace("Skipping Region size 
report to HMaster as stub is null");
-1224  return true;
-1225}
-1226try {
-1227  buildReportAndSend(rss, 
regionSizeStore);
-1228} catch (ServiceException se) {
-1229  IOException ioe = 
ProtobufUtil.getRemoteException(se);
-1230  if (ioe instanceof 
PleaseHoldException) {
-1231LOG.trace("Failed to report 
region sizes to Master because it is initializing."
-1232+ " This will be retried.", 
ioe);
-1233// The Master is coming up. Will 
retry the report later. Avoid re-creating the stub.
-1234return true;
-1235  }
-1236  if (rssStub == rss) {
-1237rssStub = null;
-1238  }
-1239  
createRegionServerStatusStub(true);
-1240  if (ioe instanceof 
DoNotRetryIOException) {
-1241DoNotRetryIOException 
doNotRetryEx = (DoNotRetryIOException) ioe;
-1242if (doNotRetryEx.getCause() != 
null) {
-1243  Throwable t = 
doNotRetryEx.getCause();
-1244  if (t instanceof 
UnsupportedOperationException) {
-1245LOG.debug("master doesn't 
support ReportRegionSpaceUse, pause before retrying");
-1246return false;
-1247  }
-1248}
-1249  }
-1250  LOG.debug("Failed to report region 
sizes to Master. This will be retried.", ioe);
-1251}
-1252return true;
-1253  }
-1254
-1255  /**
-1256   * Builds the region size report and 
sends it to the master. Upon successful sending of the
-1257   * report, the region sizes that were 
sent are marked as sent.
-1258   *
-1259   * @param rss The stub to send to the 
Master
-1260   * @param regionSizeStore The store 
containing region sizes
-1261   */
-1262  void 
buildReportAndSend(RegionServerStatusService.BlockingInterface rss,
-1263  RegionSizeStore regionSizeStore) 
throws ServiceException {
-1264RegionSpaceUseReportRequest request 
=
-1265
buildRegionSpaceUseReportRequest(Objects.requireNonNull(regionSizeStore));
-1266rss.reportRegionSpaceUse(null, 
request);
-1267// Record the number of size reports 
sent
-1268if (metricsRegionServer != null) {
-1269  
metricsRegionServer.incrementNumRegionSizeReportsSent(regionSizeStore.size());
-1270}
-1271  }
-1272
-1273  /**
-1274   * Builds a {@link 
RegionSpaceUseReportRequest} protobuf message from the region size map.
-1275   *
-1276   * @param regionSizeStore The size in 
bytes of regions
-1277   * @return The corresponding protocol 
buffer message.
-1278   */
-1279  RegionSpaceUseReportRequest 
buildRegionSpaceUseReportRequest(RegionSizeStore regionSizes) {
-1280RegionSpaceUseReportRequest.Builder 
request = 

[23/51] [partial] hbase-site git commit: Published site at cf529f18a9959589fa635f78df4840472526ea2c.

2018-05-17 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
index 8b2674f..274eb54 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
@@ -262,1413 +262,1417 @@
 254});
 255  }
 256
-257  public void preModifyNamespace(final 
NamespaceDescriptor ns) throws IOException {
-258
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-259  @Override
-260  public void call(MasterObserver 
observer) throws IOException {
-261observer.preModifyNamespace(this, 
ns);
-262  }
-263});
-264  }
-265
-266  public void postModifyNamespace(final 
NamespaceDescriptor ns) throws IOException {
-267
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-268  @Override
-269  public void call(MasterObserver 
observer) throws IOException {
-270
observer.postModifyNamespace(this, ns);
-271  }
-272});
-273  }
-274
-275  public void 
preGetNamespaceDescriptor(final String namespaceName)
-276  throws IOException {
-277
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-278  @Override
-279  public void call(MasterObserver 
observer) throws IOException {
-280
observer.preGetNamespaceDescriptor(this, namespaceName);
-281  }
-282});
-283  }
-284
-285  public void 
postGetNamespaceDescriptor(final NamespaceDescriptor ns)
-286  throws IOException {
-287
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-288  @Override
-289  public void call(MasterObserver 
observer) throws IOException {
-290
observer.postGetNamespaceDescriptor(this, ns);
-291  }
-292});
-293  }
-294
-295  public void 
preListNamespaceDescriptors(final ListNamespaceDescriptor 
descriptors)
-296  throws IOException {
-297
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-298  @Override
-299  public void call(MasterObserver 
observer) throws IOException {
-300
observer.preListNamespaceDescriptors(this, descriptors);
-301  }
-302});
-303  }
-304
-305  public void 
postListNamespaceDescriptors(final ListNamespaceDescriptor 
descriptors)
-306  throws IOException {
-307
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-308  @Override
-309  public void call(MasterObserver 
observer) throws IOException {
-310
observer.postListNamespaceDescriptors(this, descriptors);
-311  }
-312});
-313  }
-314
-315  /* Implementation of hooks for invoking 
MasterObservers */
+257  public void preModifyNamespace(final 
NamespaceDescriptor currentNsDescriptor,
+258final NamespaceDescriptor 
newNsDescriptor) throws IOException {
+259
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
+260  @Override
+261  public void call(MasterObserver 
observer) throws IOException {
+262observer.preModifyNamespace(this, 
currentNsDescriptor, newNsDescriptor);
+263  }
+264});
+265  }
+266
+267  public void postModifyNamespace(final 
NamespaceDescriptor oldNsDescriptor,
+268final NamespaceDescriptor 
currentNsDescriptor) throws IOException {
+269
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
+270  @Override
+271  public void call(MasterObserver 
observer) throws IOException {
+272
observer.postModifyNamespace(this, oldNsDescriptor, currentNsDescriptor);
+273  }
+274});
+275  }
+276
+277  public void 
preGetNamespaceDescriptor(final String namespaceName)
+278  throws IOException {
+279
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
+280  @Override
+281  public void call(MasterObserver 
observer) throws IOException {
+282
observer.preGetNamespaceDescriptor(this, namespaceName);
+283  }
+284});
+285  }
+286
+287  public void 
postGetNamespaceDescriptor(final NamespaceDescriptor ns)
+288  throws IOException {
+289
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
+290  @Override
+291  public void call(MasterObserver 
observer) throws IOException {
+292
observer.postGetNamespaceDescriptor(this, ns);
+293  }
+294});
+295  }
+296
+297  public void 
preListNamespaceDescriptors(final ListNamespaceDescriptor 
descriptors)
+298  throws IOException {
+299
execOperation(coprocEnvironments.isEmpty() ? null : new 

[23/51] [partial] hbase-site git commit: Published site at 021f66d11d2cbb7308308093e29e69d6e7661ee9.

2018-05-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
index 4a879bb..7d27402 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
@@ -300,7 +300,7 @@
 292  private MapString, 
com.google.protobuf.Service coprocessorServiceHandlers = 
Maps.newHashMap();
 293
 294  // Track data size in all memstores
-295  private final MemStoreSizing 
memStoreSize = new MemStoreSizing();
+295  private final MemStoreSizing 
memStoreSizing = new ThreadSafeMemStoreSizing();
 296  private final RegionServicesForStores 
regionServicesForStores = new RegionServicesForStores(this);
 297
 298  // Debug possible data loss due to WAL 
off
@@ -1218,7389 +1218,7399 @@
 1210   * Increase the size of mem store in 
this region and the size of global mem
 1211   * store
 1212   */
-1213  public void 
incMemStoreSize(MemStoreSize memStoreSize) {
-1214if (this.rsAccounting != null) {
-1215  
rsAccounting.incGlobalMemStoreSize(memStoreSize);
-1216}
-1217long dataSize;
-1218synchronized (this.memStoreSize) {
-1219  
this.memStoreSize.incMemStoreSize(memStoreSize);
-1220  dataSize = 
this.memStoreSize.getDataSize();
-1221}
-1222
checkNegativeMemStoreDataSize(dataSize, memStoreSize.getDataSize());
-1223  }
-1224
-1225  public void 
decrMemStoreSize(MemStoreSize memStoreSize) {
-1226if (this.rsAccounting != null) {
-1227  
rsAccounting.decGlobalMemStoreSize(memStoreSize);
-1228}
-1229long size;
-1230synchronized (this.memStoreSize) {
-1231  
this.memStoreSize.decMemStoreSize(memStoreSize);
-1232  size = 
this.memStoreSize.getDataSize();
+1213  void incMemStoreSize(MemStoreSize mss) 
{
+1214incMemStoreSize(mss.getDataSize(), 
mss.getHeapSize(), mss.getOffHeapSize());
+1215  }
+1216
+1217  void incMemStoreSize(long 
dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) {
+1218if (this.rsAccounting != null) {
+1219  
rsAccounting.incGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, 
offHeapSizeDelta);
+1220}
+1221long dataSize =
+1222
this.memStoreSizing.incMemStoreSize(dataSizeDelta, heapSizeDelta, 
offHeapSizeDelta);
+1223
checkNegativeMemStoreDataSize(dataSize, dataSizeDelta);
+1224  }
+1225
+1226  void decrMemStoreSize(MemStoreSize 
mss) {
+1227decrMemStoreSize(mss.getDataSize(), 
mss.getHeapSize(), mss.getOffHeapSize());
+1228  }
+1229
+1230  void decrMemStoreSize(long 
dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) {
+1231if (this.rsAccounting != null) {
+1232  
rsAccounting.decGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, 
offHeapSizeDelta);
 1233}
-1234checkNegativeMemStoreDataSize(size, 
-memStoreSize.getDataSize());
-1235  }
-1236
-1237  private void 
checkNegativeMemStoreDataSize(long memStoreDataSize, long delta) {
-1238// This is extremely bad if we make 
memStoreSize negative. Log as much info on the offending
-1239// caller as possible. (memStoreSize 
might be a negative value already -- freeing memory)
-1240if (memStoreDataSize  0) {
-1241  LOG.error("Asked to modify this 
region's (" + this.toString()
-1242  + ") memStoreSize to a 
negative value which is incorrect. Current memStoreSize="
-1243  + (memStoreDataSize - delta) + 
", delta=" + delta, new Exception());
-1244}
-1245  }
-1246
-1247  @Override
-1248  public RegionInfo getRegionInfo() {
-1249return this.fs.getRegionInfo();
-1250  }
-1251
-1252  /**
-1253   * @return Instance of {@link 
RegionServerServices} used by this HRegion.
-1254   * Can be null.
-1255   */
-1256  RegionServerServices 
getRegionServerServices() {
-1257return this.rsServices;
-1258  }
-1259
-1260  @Override
-1261  public long getReadRequestsCount() {
-1262return readRequestsCount.sum();
-1263  }
-1264
-1265  @Override
-1266  public long 
getFilteredReadRequestsCount() {
-1267return 
filteredReadRequestsCount.sum();
-1268  }
-1269
-1270  @Override
-1271  public long getWriteRequestsCount() 
{
-1272return writeRequestsCount.sum();
-1273  }
-1274
-1275  @Override
-1276  public long getMemStoreDataSize() {
-1277return memStoreSize.getDataSize();
-1278  }
-1279
-1280  @Override
-1281  public long getMemStoreHeapSize() {
-1282return memStoreSize.getHeapSize();
-1283  }
-1284
-1285  @Override
-1286  public long getMemStoreOffHeapSize() 
{
-1287return 
memStoreSize.getOffHeapSize();
-1288  }
-1289
-1290  /** @return store services for this 
region, to access services required by store level needs */
-1291  public RegionServicesForStores 

[23/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.

2018-05-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html
index 2510283..418c60c 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html
@@ -77,77 +77,77 @@
 069import 
org.apache.hadoop.hbase.client.RowMutations;
 070import 
org.apache.hadoop.hbase.client.Scan;
 071import 
org.apache.hadoop.hbase.client.Table;
-072import 
org.apache.hadoop.hbase.filter.BinaryComparator;
-073import 
org.apache.hadoop.hbase.filter.Filter;
-074import 
org.apache.hadoop.hbase.filter.FilterAllFilter;
-075import 
org.apache.hadoop.hbase.filter.FilterList;
-076import 
org.apache.hadoop.hbase.filter.PageFilter;
-077import 
org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
-078import 
org.apache.hadoop.hbase.filter.WhileMatchFilter;
-079import 
org.apache.hadoop.hbase.io.compress.Compression;
-080import 
org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
-081import 
org.apache.hadoop.hbase.io.hfile.RandomDistribution;
-082import 
org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
-083import 
org.apache.hadoop.hbase.regionserver.BloomType;
-084import 
org.apache.hadoop.hbase.regionserver.CompactingMemStore;
-085import 
org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration;
-086import 
org.apache.hadoop.hbase.trace.SpanReceiverHost;
-087import 
org.apache.hadoop.hbase.trace.TraceUtil;
-088import 
org.apache.hadoop.hbase.util.ByteArrayHashKey;
-089import 
org.apache.hadoop.hbase.util.Bytes;
-090import 
org.apache.hadoop.hbase.util.Hash;
-091import 
org.apache.hadoop.hbase.util.MurmurHash;
-092import 
org.apache.hadoop.hbase.util.Pair;
-093import 
org.apache.hadoop.hbase.util.YammerHistogramUtils;
-094import 
org.apache.hadoop.io.LongWritable;
-095import org.apache.hadoop.io.Text;
-096import org.apache.hadoop.mapreduce.Job;
-097import 
org.apache.hadoop.mapreduce.Mapper;
-098import 
org.apache.hadoop.mapreduce.lib.input.NLineInputFormat;
-099import 
org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
-100import 
org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer;
-101import org.apache.hadoop.util.Tool;
-102import 
org.apache.hadoop.util.ToolRunner;
-103import 
org.apache.htrace.core.ProbabilitySampler;
-104import org.apache.htrace.core.Sampler;
-105import 
org.apache.htrace.core.TraceScope;
-106import 
org.apache.yetus.audience.InterfaceAudience;
-107import org.slf4j.Logger;
-108import org.slf4j.LoggerFactory;
-109import 
org.apache.hbase.thirdparty.com.google.common.base.MoreObjects;
-110import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-111
-112/**
-113 * Script used evaluating HBase 
performance and scalability.  Runs a HBase
-114 * client that steps through one of a set 
of hardcoded tests or 'experiments'
-115 * (e.g. a random reads test, a random 
writes test, etc.). Pass on the
-116 * command-line which test to run and how 
many clients are participating in
-117 * this experiment. Run {@code 
PerformanceEvaluation --help} to obtain usage.
-118 *
-119 * pThis class sets up and runs 
the evaluation programs described in
-120 * Section 7, iPerformance 
Evaluation/i, of the a
-121 * 
href="http://labs.google.com/papers/bigtable.html"Bigtable/a;
-122 * paper, pages 8-10.
-123 *
-124 * pBy default, runs as a 
mapreduce job where each mapper runs a single test
-125 * client. Can also run as a 
non-mapreduce, multithreaded application by
-126 * specifying {@code --nomapred}. Each 
client does about 1GB of data, unless
-127 * specified otherwise.
-128 */
-129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
-130public class PerformanceEvaluation 
extends Configured implements Tool {
-131  static final String RANDOM_SEEK_SCAN = 
"randomSeekScan";
-132  static final String RANDOM_READ = 
"randomRead";
-133  private static final Logger LOG = 
LoggerFactory.getLogger(PerformanceEvaluation.class.getName());
-134  private static final ObjectMapper 
MAPPER = new ObjectMapper();
-135  static {
-136
MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true);
-137  }
-138
-139  public static final String TABLE_NAME = 
"TestTable";
-140  public static final byte[] FAMILY_NAME 
= Bytes.toBytes("info");
-141  public static final byte [] COLUMN_ZERO 
= Bytes.toBytes("" + 0);
-142  public static final byte [] 
QUALIFIER_NAME = COLUMN_ZERO;
+072import 
org.apache.hadoop.hbase.client.metrics.ScanMetrics;
+073import 
org.apache.hadoop.hbase.filter.BinaryComparator;
+074import 
org.apache.hadoop.hbase.filter.Filter;
+075import 

[23/51] [partial] hbase-site git commit: Published site at 87f5b5f3411d96c31b4cb61b9a57ced22be91d1f.

2018-05-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html
index e1bc325..63e7421 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html
@@ -66,5125 +66,5224 @@
 058import 
java.util.concurrent.TimeoutException;
 059import 
java.util.concurrent.atomic.AtomicBoolean;
 060import 
java.util.concurrent.atomic.AtomicInteger;
-061import org.apache.commons.io.IOUtils;
-062import 
org.apache.commons.lang3.RandomStringUtils;
-063import 
org.apache.commons.lang3.StringUtils;
-064import 
org.apache.hadoop.conf.Configuration;
-065import 
org.apache.hadoop.conf.Configured;
-066import 
org.apache.hadoop.fs.FSDataOutputStream;
-067import org.apache.hadoop.fs.FileStatus;
-068import org.apache.hadoop.fs.FileSystem;
-069import org.apache.hadoop.fs.Path;
-070import 
org.apache.hadoop.fs.permission.FsAction;
-071import 
org.apache.hadoop.fs.permission.FsPermission;
-072import 
org.apache.hadoop.hbase.Abortable;
-073import org.apache.hadoop.hbase.Cell;
-074import 
org.apache.hadoop.hbase.CellUtil;
-075import 
org.apache.hadoop.hbase.ClusterMetrics;
-076import 
org.apache.hadoop.hbase.ClusterMetrics.Option;
-077import 
org.apache.hadoop.hbase.HBaseConfiguration;
-078import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-079import 
org.apache.hadoop.hbase.HConstants;
-080import 
org.apache.hadoop.hbase.HRegionInfo;
-081import 
org.apache.hadoop.hbase.HRegionLocation;
-082import 
org.apache.hadoop.hbase.KeyValue;
-083import 
org.apache.hadoop.hbase.MasterNotRunningException;
-084import 
org.apache.hadoop.hbase.MetaTableAccessor;
-085import 
org.apache.hadoop.hbase.RegionLocations;
-086import 
org.apache.hadoop.hbase.ServerName;
-087import 
org.apache.hadoop.hbase.TableName;
-088import 
org.apache.hadoop.hbase.ZooKeeperConnectionException;
-089import 
org.apache.hadoop.hbase.client.Admin;
-090import 
org.apache.hadoop.hbase.client.ClusterConnection;
-091import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
-092import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
-093import 
org.apache.hadoop.hbase.client.Connection;
-094import 
org.apache.hadoop.hbase.client.ConnectionFactory;
-095import 
org.apache.hadoop.hbase.client.Delete;
-096import 
org.apache.hadoop.hbase.client.Get;
-097import 
org.apache.hadoop.hbase.client.Put;
-098import 
org.apache.hadoop.hbase.client.RegionInfo;
-099import 
org.apache.hadoop.hbase.client.RegionInfoBuilder;
-100import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-101import 
org.apache.hadoop.hbase.client.Result;
-102import 
org.apache.hadoop.hbase.client.RowMutations;
-103import 
org.apache.hadoop.hbase.client.Table;
-104import 
org.apache.hadoop.hbase.client.TableDescriptor;
-105import 
org.apache.hadoop.hbase.client.TableDescriptorBuilder;
-106import 
org.apache.hadoop.hbase.client.TableState;
-107import 
org.apache.hadoop.hbase.io.FileLink;
-108import 
org.apache.hadoop.hbase.io.HFileLink;
-109import 
org.apache.hadoop.hbase.io.hfile.CacheConfig;
-110import 
org.apache.hadoop.hbase.io.hfile.HFile;
-111import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-112import 
org.apache.hadoop.hbase.master.MasterFileSystem;
-113import 
org.apache.hadoop.hbase.master.RegionState;
-114import 
org.apache.hadoop.hbase.regionserver.HRegion;
-115import 
org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
-116import 
org.apache.hadoop.hbase.regionserver.StoreFileInfo;
-117import 
org.apache.hadoop.hbase.replication.ReplicationException;
-118import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-119import 
org.apache.hadoop.hbase.security.UserProvider;
-120import 
org.apache.hadoop.hbase.util.Bytes.ByteArrayComparator;
-121import 
org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE;
-122import 
org.apache.hadoop.hbase.util.hbck.HFileCorruptionChecker;
-123import 
org.apache.hadoop.hbase.util.hbck.ReplicationChecker;
-124import 
org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandler;
-125import 
org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandlerImpl;
-126import org.apache.hadoop.hbase.wal.WAL;
-127import 
org.apache.hadoop.hbase.wal.WALFactory;
-128import 
org.apache.hadoop.hbase.wal.WALSplitter;
-129import 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator;
-130import 
org.apache.hadoop.hbase.zookeeper.ZKUtil;
-131import 
org.apache.hadoop.hbase.zookeeper.ZKWatcher;
-132import 
org.apache.hadoop.hbase.zookeeper.ZNodePaths;
-133import 
org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException;
-134import 
org.apache.hadoop.ipc.RemoteException;
-135import 

[23/51] [partial] hbase-site git commit: Published site at 2912c953551bedbfbf30c32c156ed7bb187d54c3.

2018-04-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html
index e6e43ee..a8b77ae 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html
@@ -367,650 +367,650 @@
 359lock.lock();
 360try {
 361  LOG.trace("Starting WAL Procedure 
Store lease recovery");
-362  FileStatus[] oldLogs = 
getLogFiles();
-363  while (isRunning()) {
+362  while (isRunning()) {
+363FileStatus[] oldLogs = 
getLogFiles();
 364// Get Log-MaxID and recover 
lease on old logs
 365try {
 366  flushLogId = 
initOldLogs(oldLogs);
 367} catch (FileNotFoundException e) 
{
 368  LOG.warn("Someone else is 
active and deleted logs. retrying.", e);
-369  oldLogs = getLogFiles();
-370  continue;
-371}
-372
-373// Create new state-log
-374if (!rollWriter(flushLogId + 1)) 
{
-375  // someone else has already 
created this log
-376  LOG.debug("Someone else has 
already created log " + flushLogId);
-377  continue;
-378}
-379
-380// We have the lease on the log
-381oldLogs = getLogFiles();
-382if (getMaxLogId(oldLogs)  
flushLogId) {
-383  if (LOG.isDebugEnabled()) {
-384LOG.debug("Someone else 
created new logs. Expected maxLogId  " + flushLogId);
-385  }
-386  
logs.getLast().removeFile(this.walArchiveDir);
-387  continue;
-388}
-389
-390LOG.trace("Lease acquired for 
flushLogId={}", flushLogId);
-391break;
-392  }
-393} finally {
-394  lock.unlock();
-395}
-396  }
-397
-398  @Override
-399  public void load(final ProcedureLoader 
loader) throws IOException {
-400lock.lock();
-401try {
-402  if (logs.isEmpty()) {
-403throw new 
RuntimeException("recoverLease() must be called before loading data");
-404  }
-405
-406  // Nothing to do, If we have only 
the current log.
-407  if (logs.size() == 1) {
-408LOG.trace("No state logs to 
replay.");
-409loader.setMaxProcId(0);
-410return;
-411  }
-412
-413  // Load the old logs
-414  final 
IteratorProcedureWALFile it = logs.descendingIterator();
-415  it.next(); // Skip the current 
log
-416
-417  ProcedureWALFormat.load(it, 
storeTracker, new ProcedureWALFormat.Loader() {
-418@Override
-419public void setMaxProcId(long 
maxProcId) {
-420  
loader.setMaxProcId(maxProcId);
-421}
-422
-423@Override
-424public void 
load(ProcedureIterator procIter) throws IOException {
-425  loader.load(procIter);
-426}
-427
-428@Override
-429public void 
handleCorrupted(ProcedureIterator procIter) throws IOException {
-430  
loader.handleCorrupted(procIter);
-431}
-432
-433@Override
-434public void 
markCorruptedWAL(ProcedureWALFile log, IOException e) {
-435  if (corruptedLogs == null) {
-436corruptedLogs = new 
HashSet();
-437  }
-438  corruptedLogs.add(log);
-439  // TODO: sideline corrupted 
log
-440}
-441  });
-442} finally {
-443  try {
-444// try to cleanup inactive wals 
and complete the operation
-445buildHoldingCleanupTracker();
-446tryCleanupLogsOnLoad();
-447loading.set(false);
-448  } finally {
-449lock.unlock();
-450  }
-451}
-452  }
-453
-454  private void tryCleanupLogsOnLoad() {
-455// nothing to cleanup.
-456if (logs.size() = 1) return;
-457
-458// the config says to not cleanup 
wals on load.
-459if 
(!conf.getBoolean(EXEC_WAL_CLEANUP_ON_LOAD_CONF_KEY,
-460  
DEFAULT_EXEC_WAL_CLEANUP_ON_LOAD_CONF_KEY)) {
-461  LOG.debug("WALs cleanup on load is 
not enabled: " + getActiveLogs());
-462  return;
-463}
-464
-465try {
-466  periodicRoll();
-467} catch (IOException e) {
-468  LOG.warn("Unable to cleanup logs on 
load: " + e.getMessage(), e);
-469}
-470  }
-471
-472  @Override
-473  public void insert(final Procedure 
proc, final Procedure[] subprocs) {
-474if (LOG.isTraceEnabled()) {
-475  LOG.trace("Insert " + proc + ", 
subproc=" + Arrays.toString(subprocs));
-476}
-477
-478ByteSlot slot = acquireSlot();
-479try {
-480  // Serialize the insert
-481  long[] subProcIds = null;
-482  if (subprocs != null) {
-483

[23/51] [partial] hbase-site git commit: Published site at 2a2258656b2fcd92b967131b6c1f037363553bc4.

2018-03-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html
index 1038256..89735e7 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html
@@ -36,398 +36,399 @@
 028import java.util.List;
 029import java.util.Map;
 030
-031import 
org.apache.commons.collections4.iterators.UnmodifiableIterator;
-032import 
org.apache.hadoop.conf.Configuration;
-033import 
org.apache.hadoop.hbase.util.Bytes;
-034import 
org.apache.yetus.audience.InterfaceAudience;
-035
-036/**
-037 * Do a shallow merge of multiple KV 
configuration pools. This is a very useful
-038 * utility class to easily add per-object 
configurations in addition to wider
-039 * scope settings. This is different from 
Configuration.addResource()
-040 * functionality, which performs a deep 
merge and mutates the common data
-041 * structure.
-042 * p
-043 * The iterator on CompoundConfiguration 
is unmodifiable. Obtaining iterator is an expensive
-044 * operation.
-045 * p
-046 * For clarity: the shallow merge allows 
the user to mutate either of the
-047 * configuration objects and have changes 
reflected everywhere. In contrast to a
-048 * deep merge, that requires you to 
explicitly know all applicable copies to
-049 * propagate changes.
-050 * 
-051 * WARNING: The values set in the 
CompoundConfiguration are do not handle Property variable
-052 * substitution.  However, if they are 
set in the underlying configuration substitutions are
-053 * done. 
-054 */
-055@InterfaceAudience.Private
-056public class CompoundConfiguration 
extends Configuration {
-057
-058  private Configuration mutableConf = 
null;
-059
-060  /**
-061   * Default Constructor. Initializes 
empty configuration
-062   */
-063  public CompoundConfiguration() {
-064  }
-065
-066  // Devs: these APIs are the same 
contract as their counterparts in
-067  // Configuration.java
-068  private interface ImmutableConfigMap 
extends IterableMap.EntryString,String {
-069String get(String key);
-070String getRaw(String key);
-071Class? getClassByName(String 
name) throws ClassNotFoundException;
-072int size();
-073  }
-074
-075  private final 
ListImmutableConfigMap configs = new ArrayList();
-076
-077  static class ImmutableConfWrapper 
implements  ImmutableConfigMap {
-078   private final Configuration c;
-079
-080ImmutableConfWrapper(Configuration 
conf) {
-081  c = conf;
-082}
-083
-084@Override
-085public 
IteratorMap.EntryString,String iterator() {
-086  return c.iterator();
-087}
-088
-089@Override
-090public String get(String key) {
-091  return c.get(key);
-092}
-093
-094@Override
-095public String getRaw(String key) {
-096  return c.getRaw(key);
-097}
-098
-099@Override
-100public Class? 
getClassByName(String name)
-101throws ClassNotFoundException {
-102  return c.getClassByName(name);
-103}
-104
-105@Override
-106public int size() {
-107  return c.size();
-108}
-109
-110@Override
-111public String toString() {
-112  return c.toString();
-113}
-114  }
-115
-116  /**
-117   * If set has been called, it will 
create a mutableConf.  This converts the mutableConf to an
-118   * immutable one and resets it to allow 
a new mutable conf.  This is used when a new map or
-119   * conf is added to the compound 
configuration to preserve proper override semantics.
-120   */
-121  void freezeMutableConf() {
-122if (mutableConf == null) {
-123  // do nothing if there is no 
current mutableConf
-124  return;
-125}
-126
-127this.configs.add(0, new 
ImmutableConfWrapper(mutableConf));
-128mutableConf = null;
-129  }
-130
-131  /**
-132   * Add Hadoop Configuration object to 
config list.
-133   * The added configuration overrides 
the previous ones if there are name collisions.
-134   * @param conf configuration object
-135   * @return this, for builder pattern
-136   */
-137  public CompoundConfiguration add(final 
Configuration conf) {
-138freezeMutableConf();
-139
-140if (conf instanceof 
CompoundConfiguration) {
-141  this.configs.addAll(0, 
((CompoundConfiguration) conf).configs);
-142  return this;
-143}
-144// put new config at the front of the 
list (top priority)
-145this.configs.add(0, new 
ImmutableConfWrapper(conf));
-146return this;
-147  }
-148
-149  /**
-150   * Add Bytes map to config list. This 
map is generally
-151   * created by HTableDescriptor or 
HColumnDescriptor, but can be abstractly
-152   * used. The added configuration 
overrides the previous ones if there are
-153   * name 

[23/51] [partial] hbase-site git commit: Published site at e468b4022f76688851b3e0c34722f01a56bd624f.

2018-03-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index 52f3dab..213814f 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
 en-us
 2007 - 2018 The Apache Software Foundation
 
-  File: 3592,
- Errors: 16020,
+  File: 3593,
+ Errors: 15918,
  Warnings: 0,
  Infos: 0
   
@@ -1656,7 +1656,7 @@ under the License.
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter.java;>org/apache/hadoop/hbase/filter/TestMultiRowRangeFilter.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.client.TestResult.java;>org/apache/hadoop/hbase/client/TestResult.java
 
 
   0
@@ -1665,12 +1665,12 @@ under the License.
   0
 
 
-  2
+  3
 
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.client.TestResult.java;>org/apache/hadoop/hbase/client/TestResult.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter.java;>org/apache/hadoop/hbase/filter/TestMultiRowRangeFilter.java
 
 
   0
@@ -1679,7 +1679,7 @@ under the License.
   0
 
 
-  3
+  2
 
   
   
@@ -1894,7 +1894,7 @@ under the License.
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
 
 
   0
@@ -1903,12 +1903,12 @@ under the License.
   0
 
 
-  15
+  1
 
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
 
 
   0
@@ -1917,7 +1917,7 @@ under the License.
   0
 
 
-  1
+  15
 
   
   
@@ -2286,7 +2286,7 @@ under the License.
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.java;>org/apache/hadoop/hbase/coprocessor/TestCoprocessorMetrics.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks.java;>org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.java
 
 
   0
@@ -2295,12 +2295,12 @@ under the License.
   0
 
 
-  1
+  3
 
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks.java;>org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.java;>org/apache/hadoop/hbase/coprocessor/TestCoprocessorMetrics.java
 
 
   0
@@ -2309,7 +2309,7 @@ under the License.
   0
 
 
-  3
+  1
 
   
   
@@ -3686,7 +3686,7 @@ under the License.
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.RegionServerRunningException.java;>org/apache/hadoop/hbase/regionserver/RegionServerRunningException.java
+  

[23/51] [partial] hbase-site git commit: Published site at 64061f896fe21512504e3886a400759e88b519da.

2018-03-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
index c15c700..fa49042 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
@@ -511,13 +511,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 ServerNameserverName)
 
 
+static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListPairhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long
+MetaTableAccessor.getTableEncodedRegionNameAndLastBarrier(Connectionconn,
+   TableNametableName)
+
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
 MetaTableAccessor.getTableRegions(Connectionconnection,
TableNametableName)
 Gets all of the regions of the specified table.
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
 MetaTableAccessor.getTableRegions(Connectionconnection,
TableNametableName,
@@ -525,14 +530,14 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 Gets all of the regions of the specified table.
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListPairRegionInfo,ServerName
 MetaTableAccessor.getTableRegionsAndLocations(Connectionconnection,
TableNametableName)
 Do not use this method to get meta table regions, use 
methods in MetaTableLocator instead.
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListPairRegionInfo,ServerName
 MetaTableAccessor.getTableRegionsAndLocations(Connectionconnection,
TableNametableName,
@@ -540,20 +545,20 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 Do not use this method to get meta table regions, use 
methods in MetaTableLocator instead.
 
 
-
+
 static TableState
 MetaTableAccessor.getTableState(Connectionconn,
  TableNametableName)
 Fetch table state for given table from META table
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">MapTableName,TableState
 MetaTableAccessor.getTableStates(Connectionconn)
 Fetch table states from META table
 
 
-
+
 static void
 MetaTableAccessor.mergeRegions(Connectionconnection,
 RegionInfomergedRegion,
@@ -566,7 +571,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 Merge the two regions into one in an atomic operation.
 
 
-
+
 private static void
 MetaTableAccessor.multiMutate(Connectionconnection,
Tabletable,
@@ -575,21 +580,21 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 Performs an atomic multi-mutate operation against the given 
table.
 
 
-
+
 private static void
 MetaTableAccessor.multiMutate(Connectionconnection,
Tabletable,
byte[]row,
Mutation...mutations)
 
-
+
 static void
 MetaTableAccessor.mutateMetaTable(Connectionconnection,
https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListMutationmutations)
 Execute the passed mutations against 
hbase:meta table.
 
 
-
+
 static void
 MetaTableAccessor.overwriteRegions(Connectionconnection,
 https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInforegionInfos,
@@ -597,21 +602,21 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 Overwrites the specified regions from hbase:meta.
 
 
-
+
 static void
 MetaTableAccessor.putsToMetaTable(Connectionconnection,
https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListPutps)
 Put the passed ps to the 
hbase:meta table.
 
 
-
+
 private static void
 MetaTableAccessor.putToMetaTable(Connectionconnection,
   Putp)
 Put the passed p to the 
hbase:meta table.
 
 
-
+
 static void
 

[23/51] [partial] hbase-site git commit: Published site at 4cb40e6d846ce1f28ffb40d388c9efb753197813.

2018-03-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
index d5a3666..d3a9f2d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
@@ -320,8 +320,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncHBaseAdmin.addColumnFamily(TableNametableName,
-   ColumnFamilyDescriptorcolumnFamily)
+AsyncAdmin.addColumnFamily(TableNametableName,
+   ColumnFamilyDescriptorcolumnFamily)
+Add a column family to an existing table.
+
 
 
 void
@@ -331,20 +333,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-void
-HBaseAdmin.addColumnFamily(TableNametableName,
+https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
+RawAsyncHBaseAdmin.addColumnFamily(TableNametableName,
ColumnFamilyDescriptorcolumnFamily)
 
 
-https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncAdmin.addColumnFamily(TableNametableName,
-   ColumnFamilyDescriptorcolumnFamily)
-Add a column family to an existing table.
-
+void
+HBaseAdmin.addColumnFamily(TableNametableName,
+   ColumnFamilyDescriptorcolumnFamily)
 
 
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-RawAsyncHBaseAdmin.addColumnFamily(TableNametableName,
+AsyncHBaseAdmin.addColumnFamily(TableNametableName,
ColumnFamilyDescriptorcolumnFamily)
 
 
@@ -386,8 +386,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncHBaseAdmin.modifyColumnFamily(TableNametableName,
-  ColumnFamilyDescriptorcolumnFamily)
+AsyncAdmin.modifyColumnFamily(TableNametableName,
+  ColumnFamilyDescriptorcolumnFamily)
+Modify an existing column family on a table.
+
 
 
 void
@@ -397,20 +399,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-void
-HBaseAdmin.modifyColumnFamily(TableNametableName,
+https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
+RawAsyncHBaseAdmin.modifyColumnFamily(TableNametableName,
   ColumnFamilyDescriptorcolumnFamily)
 
 
-https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncAdmin.modifyColumnFamily(TableNametableName,
-  ColumnFamilyDescriptorcolumnFamily)
-Modify an existing column family on a table.
-
+void
+HBaseAdmin.modifyColumnFamily(TableNametableName,
+  ColumnFamilyDescriptorcolumnFamily)
 
 
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void

[23/51] [partial] hbase-site git commit: Published site at 8ab7b20f48951d77945181024f5e15842bc253c4.

2018-03-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
index ecf500c..0cd5a4e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
@@ -238,8355 +238,8368 @@
 230  public static final String 
HBASE_MAX_CELL_SIZE_KEY = "hbase.server.keyvalue.maxsize";
 231  public static final int 
DEFAULT_MAX_CELL_SIZE = 10485760;
 232
-233  public static final String 
HBASE_REGIONSERVER_MINIBATCH_SIZE =
-234  
"hbase.regionserver.minibatch.size";
-235  public static final int 
DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2;
-236
-237  /**
-238   * This is the global default value for 
durability. All tables/mutations not
-239   * defining a durability or using 
USE_DEFAULT will default to this value.
-240   */
-241  private static final Durability 
DEFAULT_DURABILITY = Durability.SYNC_WAL;
+233  /**
+234   * This is the global default value for 
durability. All tables/mutations not
+235   * defining a durability or using 
USE_DEFAULT will default to this value.
+236   */
+237  private static final Durability 
DEFAULT_DURABILITY = Durability.SYNC_WAL;
+238
+239  public static final String 
HBASE_REGIONSERVER_MINIBATCH_SIZE =
+240  
"hbase.regionserver.minibatch.size";
+241  public static final int 
DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2;
 242
-243  final AtomicBoolean closed = new 
AtomicBoolean(false);
-244
-245  /* Closing can take some time; use the 
closing flag if there is stuff we don't
-246   * want to do while in closing state; 
e.g. like offer this region up to the
-247   * master as a region to close if the 
carrying regionserver is overloaded.
-248   * Once set, it is never cleared.
-249   */
-250  final AtomicBoolean closing = new 
AtomicBoolean(false);
-251
-252  /**
-253   * The max sequence id of flushed data 
on this region. There is no edit in memory that is
-254   * less that this sequence id.
-255   */
-256  private volatile long maxFlushedSeqId = 
HConstants.NO_SEQNUM;
-257
-258  /**
-259   * Record the sequence id of last flush 
operation. Can be in advance of
-260   * {@link #maxFlushedSeqId} when 
flushing a single column family. In this case,
-261   * {@link #maxFlushedSeqId} will be 
older than the oldest edit in memory.
-262   */
-263  private volatile long lastFlushOpSeqId 
= HConstants.NO_SEQNUM;
-264
-265  /**
-266   * The sequence id of the last replayed 
open region event from the primary region. This is used
-267   * to skip entries before this due to 
the possibility of replay edits coming out of order from
-268   * replication.
-269   */
-270  protected volatile long 
lastReplayedOpenRegionSeqId = -1L;
-271  protected volatile long 
lastReplayedCompactionSeqId = -1L;
-272
-273  
//
-274  // Members
-275  
//
-276
-277  // map from a locked row to the context 
for that lock including:
-278  // - CountDownLatch for threads waiting 
on that row
-279  // - the thread that owns the lock 
(allow reentrancy)
-280  // - reference count of (reentrant) 
locks held by the thread
-281  // - the row itself
-282  private final 
ConcurrentHashMapHashedBytes, RowLockContext lockedRows =
-283  new ConcurrentHashMap();
-284
-285  protected final Mapbyte[], 
HStore stores =
-286  new 
ConcurrentSkipListMap(Bytes.BYTES_RAWCOMPARATOR);
+243  public static final String 
WAL_HSYNC_CONF_KEY = "hbase.wal.hsync";
+244  public static final boolean 
DEFAULT_WAL_HSYNC = false;
+245
+246  final AtomicBoolean closed = new 
AtomicBoolean(false);
+247
+248  /* Closing can take some time; use the 
closing flag if there is stuff we don't
+249   * want to do while in closing state; 
e.g. like offer this region up to the
+250   * master as a region to close if the 
carrying regionserver is overloaded.
+251   * Once set, it is never cleared.
+252   */
+253  final AtomicBoolean closing = new 
AtomicBoolean(false);
+254
+255  /**
+256   * The max sequence id of flushed data 
on this region. There is no edit in memory that is
+257   * less that this sequence id.
+258   */
+259  private volatile long maxFlushedSeqId = 
HConstants.NO_SEQNUM;
+260
+261  /**
+262   * Record the sequence id of last flush 
operation. Can be in advance of
+263   * {@link #maxFlushedSeqId} when 
flushing a single column family. In this case,
+264   * {@link #maxFlushedSeqId} will be 
older than the oldest edit in memory.
+265   */
+266  private volatile long 

[23/51] [partial] hbase-site git commit: Published site at 00095a2ef9442e3fd86c04876c9d91f2f8b23ad8.

2018-03-17 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html
index c27b109..4160a88 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html
@@ -105,7 +105,7 @@
 097try {
 098  done = waitUntilDone(startTime 
* 1000L + asyncProcess.primaryCallTimeoutMicroseconds);
 099} catch (InterruptedException ex) 
{
-100  LOG.error("Replica thread was 
interrupted - no replica calls: " + ex.getMessage());
+100  LOG.error("Replica thread 
interrupted - no replica calls {}", ex.getMessage());
 101  return;
 102}
 103  }
@@ -149,7 +149,7 @@
 141  if (loc == null) return;
 142  HRegionLocation[] locs = 
loc.getRegionLocations();
 143  if (locs.length == 1) {
-144LOG.warn("No replicas found for " 
+ action.getAction());
+144LOG.warn("No replicas found for 
{}", action.getAction());
 145return;
 146  }
 147  synchronized (replicaResultLock) 
{
@@ -230,8 +230,8 @@
 222  return;
 223} catch (Throwable t) {
 224  // This should not happen. 
Let's log  retry anyway.
-225  LOG.error("#" + asyncProcess.id 
+ ", Caught throwable while calling. This is unexpected." +
-226  " Retrying. Server is " + 
server + ", tableName=" + tableName, t);
+225  LOG.error("id=" + 
asyncProcess.id + ", caught throwable. Unexpected." +
+226  " Retrying. Server=" + 
server + ", tableName=" + tableName, t);
 227  
receiveGlobalFailure(multiAction, server, numAttempt, t);
 228  return;
 229}
@@ -247,1036 +247,1035 @@
 239}
 240  } catch (Throwable t) {
 241// Something really bad happened. 
We are on the send thread that will now die.
-242LOG.error("Internal AsyncProcess 
#" + asyncProcess.id + " error for "
-243+ tableName + " processing 
for " + server, t);
-244throw new RuntimeException(t);
-245  } finally {
-246
asyncProcess.decTaskCounters(multiAction.getRegions(), server);
-247if (callsInProgress != null 
 callable != null  res != null) {
-248  
callsInProgress.remove(callable);
-249}
-250  }
-251}
-252  }
-253
-254  private final 
Batch.CallbackCResult callback;
-255  private final BatchErrors errors;
-256  private final 
ConnectionImplementation.ServerErrorTracker errorsByServer;
-257  private final ExecutorService pool;
-258  private final 
SetCancellableRegionServerCallable callsInProgress;
+242LOG.error("id=" + asyncProcess.id 
+ " error for " + tableName + " processing " + server, t);
+243throw new RuntimeException(t);
+244  } finally {
+245
asyncProcess.decTaskCounters(multiAction.getRegions(), server);
+246if (callsInProgress != null 
 callable != null  res != null) {
+247  
callsInProgress.remove(callable);
+248}
+249  }
+250}
+251  }
+252
+253  private final 
Batch.CallbackCResult callback;
+254  private final BatchErrors errors;
+255  private final 
ConnectionImplementation.ServerErrorTracker errorsByServer;
+256  private final ExecutorService pool;
+257  private final 
SetCancellableRegionServerCallable callsInProgress;
+258
 259
-260
-261  private final TableName tableName;
-262  private final AtomicLong 
actionsInProgress = new AtomicLong(-1);
-263  /**
-264   * The lock controls access to results. 
It is only held when populating results where
-265   * there might be several callers 
(eventual consistency gets). For other requests,
-266   * there's one unique call going on per 
result index.
-267   */
-268  private final Object replicaResultLock 
= new Object();
-269  /**
-270   * Result array.  Null if results are 
not needed. Otherwise, each index corresponds to
-271   * the action index in initial actions 
submitted. For most request types, has null-s for
-272   * requests that are not done, and 
result/exception for those that are done.
-273   * For eventual-consistency gets, 
initially the same applies; at some point, replica calls
-274   * might be started, and 
ReplicaResultState is put at the corresponding indices. The
-275   * returning calls check the type to 
detect when this is the case. After all calls are done,
-276   * ReplicaResultState-s are replaced 
with results for the user.
-277   */
-278  private final Object[] results;
-279  /**
-280   * Indices of replica gets in results. 
If null, all or no actions are replica-gets.
-281   */
-282  private final int[] 
replicaGetIndices;
-283  private final boolean 

[23/51] [partial] hbase-site git commit: Published site at 22f4def942f8a3367d0ca6598317e9b9a7d0cfcd.

2018-03-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html
index 26b9cc0..6fa64ff 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html
@@ -949,394 +949,411 @@
 941*/
 942  public static final float 
HBASE_CLUSTER_MINIMUM_MEMORY_THRESHOLD = 0.2f;
 943
-944  public static final Pattern 
CP_HTD_ATTR_KEY_PATTERN =
-945  
Pattern.compile("^coprocessor\\$([0-9]+)$", Pattern.CASE_INSENSITIVE);
-946
-947  /**
-948   * pre
-949   * Pattern that matches a coprocessor 
specification. Form is:
-950   * {@code coprocessor jar file 
location '|' class name ['|' priority ['|' 
arguments]]}
-951   * where arguments are {@code 
KEY '=' VALUE [,...]}
-952   * For example: {@code 
hdfs:///foo.jar|com.foo.FooRegionObserver|1001|arg1=1,arg2=2}
-953   * /pre
-954   */
-955  public static final Pattern 
CP_HTD_ATTR_VALUE_PATTERN =
-956  
Pattern.compile("(^[^\\|]*)\\|([^\\|]+)\\|[\\s]*([\\d]*)[\\s]*(\\|.*)?$");
-957
-958  public static final String 
CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN = "[^=,]+";
-959  public static final String 
CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN = "[^,]+";
-960  public static final Pattern 
CP_HTD_ATTR_VALUE_PARAM_PATTERN = Pattern.compile(
-961  "(" + 
CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN + ")=(" +
-962  
CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN + "),?");
-963  public static final String 
CP_HTD_ATTR_INCLUSION_KEY =
-964  
"hbase.coprocessor.classloader.included.classes";
-965
-966  /** The delay when re-trying a socket 
operation in a loop (HBASE-4712) */
-967  public static final int 
SOCKET_RETRY_WAIT_MS = 200;
-968
-969  /** Host name of the local machine */
-970  public static final String LOCALHOST = 
"localhost";
-971
-972  /**
-973   * If this parameter is set to true, 
then hbase will read
-974   * data and then verify checksums. 
Checksum verification
-975   * inside hdfs will be switched off.  
However, if the hbase-checksum
-976   * verification fails, then it will 
switch back to using
-977   * hdfs checksums for verifiying data 
that is being read from storage.
-978   *
-979   * If this parameter is set to false, 
then hbase will not
-980   * verify any checksums, instead it 
will depend on checksum verification
-981   * being done in the hdfs client.
-982   */
-983  public static final String 
HBASE_CHECKSUM_VERIFICATION =
-984  
"hbase.regionserver.checksum.verify";
+944  /**
+945   * @deprecated  It is used internally. 
As of release 2.0.0, this will be removed in HBase 3.0.0.
+946   */
+947  @Deprecated
+948  public static final Pattern 
CP_HTD_ATTR_KEY_PATTERN =
+949  
Pattern.compile("^coprocessor\\$([0-9]+)$", Pattern.CASE_INSENSITIVE);
+950
+951  /**
+952   * pre
+953   * Pattern that matches a coprocessor 
specification. Form is:
+954   * {@code coprocessor jar file 
location '|' class name ['|' priority ['|' 
arguments]]}
+955   * where arguments are {@code 
KEY '=' VALUE [,...]}
+956   * For example: {@code 
hdfs:///foo.jar|com.foo.FooRegionObserver|1001|arg1=1,arg2=2}
+957   * /pre
+958   * @deprecated  It is used internally. 
As of release 2.0.0, this will be removed in HBase 3.0.0.
+959   */
+960  @Deprecated
+961  public static final Pattern 
CP_HTD_ATTR_VALUE_PATTERN =
+962  
Pattern.compile("(^[^\\|]*)\\|([^\\|]+)\\|[\\s]*([\\d]*)[\\s]*(\\|.*)?$");
+963  /**
+964   * @deprecated  It is used internally. 
As of release 2.0.0, this will be removed in HBase 3.0.0.
+965   */
+966  @Deprecated
+967  public static final String 
CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN = "[^=,]+";
+968  /**
+969   * @deprecated  It is used internally. 
As of release 2.0.0, this will be removed in HBase 3.0.0.
+970   */
+971  @Deprecated
+972  public static final String 
CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN = "[^,]+";
+973  /**
+974   * @deprecated  It is used internally. 
As of release 2.0.0, this will be removed in HBase 3.0.0.
+975   */
+976  @Deprecated
+977  public static final Pattern 
CP_HTD_ATTR_VALUE_PARAM_PATTERN = Pattern.compile(
+978  "(" + 
CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN + ")=(" +
+979  
CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN + "),?");
+980  public static final String 
CP_HTD_ATTR_INCLUSION_KEY =
+981  
"hbase.coprocessor.classloader.included.classes";
+982
+983  /** The delay when re-trying a socket 
operation in a loop (HBASE-4712) */
+984  public static final int 
SOCKET_RETRY_WAIT_MS = 200;
 985
-986  public static final String LOCALHOST_IP 
= "127.0.0.1";
-987
-988  public static final String 
REGION_SERVER_HANDLER_COUNT = "hbase.regionserver.handler.count";
-989  public static final 

[23/51] [partial] hbase-site git commit: Published site at 31da4d0bce69b3a47066a5df675756087ce4dc60.

2018-03-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html
index 9ae4f41..5950b5e 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html
@@ -129,7 +129,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class ModifyTableProcedure
+public class ModifyTableProcedure
 extends AbstractStateMachineTableProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableState
 
 
@@ -400,7 +400,7 @@ extends AbstractStateMachineTableProcedure
-acquireLock,
 checkTableModifiable,
 getRegionDir,
 getUser,
 releaseLock,
 <
 a 
href="../../../../../../org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html#releaseSyncLatch--">releaseSyncLatch,
 setUser,
 toStringClassDetails
+acquireLock,
 checkTableModifiable,
 getRegionDir,
 getUser,
 preflightChecks, releaseLock,
 releaseSyncLatch,
 setUser,
 toStringClassDetails
 
 
 
@@ -443,7 +443,7 @@ extends 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
@@ -452,7 +452,7 @@ extends 
 
 unmodifiedTableDescriptor
-privateTableDescriptor unmodifiedTableDescriptor
+privateTableDescriptor unmodifiedTableDescriptor
 
 
 
@@ -461,7 +461,7 @@ extends 
 
 modifiedTableDescriptor
-privateTableDescriptor modifiedTableDescriptor
+privateTableDescriptor modifiedTableDescriptor
 
 
 
@@ -470,7 +470,7 @@ extends 
 
 deleteColumnFamilyInModify
-privateboolean deleteColumnFamilyInModify
+privateboolean deleteColumnFamilyInModify
 
 
 
@@ -479,7 +479,7 @@ extends 
 
 regionInfoList
-privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo regionInfoList
+privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo regionInfoList
 
 
 
@@ -488,7 +488,7 @@ extends 
 
 traceEnabled
-privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
+privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
 
 
 
@@ -505,7 +505,7 @@ extends 
 
 ModifyTableProcedure
-publicModifyTableProcedure()
+publicModifyTableProcedure()
 
 
 
@@ -514,8 +514,13 @@ extends 
 
 ModifyTableProcedure
-publicModifyTableProcedure(MasterProcedureEnvenv,
-TableDescriptorhtd)
+publicModifyTableProcedure(MasterProcedureEnvenv,
+TableDescriptorhtd)
+ throws HBaseIOException
+
+Throws:
+HBaseIOException
+
 
 
 
@@ -524,9 +529,14 @@ extends 
 
 ModifyTableProcedure
-publicModifyTableProcedure(MasterProcedureEnvenv,
+publicModifyTableProcedure(MasterProcedureEnvenv,
 TableDescriptorhtd,
-ProcedurePrepareLatchlatch)
+ProcedurePrepareLatchlatch)
+ throws HBaseIOException
+
+Throws:
+HBaseIOException
+
 
 
 
@@ -543,7 +553,7 @@ extends 
 
 initilize
-privatevoidinitilize()
+privatevoidinitilize()
 
 
 
@@ -552,7 +562,7 @@ extends 
 
 executeFromState
-protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv,
+protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv,
   
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)
throws https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
 title="class or interface in java.lang">InterruptedException
 Description copied from 
class:StateMachineProcedure
@@ -575,7 +585,7 @@ extends 
 
 rollbackState
-protectedvoidrollbackState(MasterProcedureEnvenv,
+protectedvoidrollbackState(MasterProcedureEnvenv,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Description copied from 
class:StateMachineProcedure
@@ -595,7 +605,7 @@ extends 
 
 isRollbackSupported
-protectedbooleanisRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)

[23/51] [partial] hbase-site git commit: Published site at 6b77786dfc46d25ac5bb5f1c8a4a9eb47b52a604.

2018-03-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
index 8225386..5955f86 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
@@ -495,7 +495,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 private static HRegionLocation
-MetaTableAccessor.getRegionLocation(Resultr,
+AsyncMetaTableAccessor.getRegionLocation(Resultr,
  RegionInforegionInfo,
  intreplicaId)
 Returns the HRegionLocation parsed from the given meta row 
Result
@@ -504,7 +504,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 private static HRegionLocation
-AsyncMetaTableAccessor.getRegionLocation(Resultr,
+MetaTableAccessor.getRegionLocation(Resultr,
  RegionInforegionInfo,
  intreplicaId)
 Returns the HRegionLocation parsed from the given meta row 
Result
@@ -941,9 +941,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-AsyncAdmin.getRegions(ServerNameserverName)
-Get all the online regions on a region server.
-
+AsyncHBaseAdmin.getRegions(ServerNameserverName)
 
 
 https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
@@ -952,22 +950,22 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
-
-
 https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
 HBaseAdmin.getRegions(ServerNamesn)
 
+
+https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
+AsyncAdmin.getRegions(ServerNameserverName)
+Get all the online regions on a region server.
+
+
 
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-AsyncHBaseAdmin.getRegions(ServerNameserverName)
+RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
 
 
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-AsyncAdmin.getRegions(TableNametableName)
-Get the regions of a given table.
-
+AsyncHBaseAdmin.getRegions(TableNametableName)
 
 
 https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
@@ -976,16 +974,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-RawAsyncHBaseAdmin.getRegions(TableNametableName)
-
-
 https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
 HBaseAdmin.getRegions(TableNametableName)
 
+
+https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo

[23/51] [partial] hbase-site git commit: Published site at 1384da71375427b522b09f06862bb5d629cef52f.

2018-03-06 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
index 253a995..f484fd7 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
@@ -798,23 +798,23 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 private Connection
-RestoreTablesClient.conn
+BackupAdminImpl.conn
 
 
-protected Connection
-TableBackupClient.conn
-
-
 (package private) Connection
 BackupCommands.Command.conn
 
+
+private Connection
+RestoreTablesClient.conn
+
 
 protected Connection
-BackupManager.conn
+TableBackupClient.conn
 
 
-private Connection
-BackupAdminImpl.conn
+protected Connection
+BackupManager.conn
 
 
 private Connection
@@ -1128,13 +1128,13 @@ Input/OutputFormats, a table indexing MapReduce job, 
and utility methods.
 
 
 
-(package private) Connection
-ConnectionImplementation.MasterServiceState.connection
-
-
 private Connection
 RegionServerCallable.connection
 
+
+(package private) Connection
+ConnectionImplementation.MasterServiceState.connection
+
 
 
 
@@ -1179,20 +1179,20 @@ Input/OutputFormats, a table indexing MapReduce job, 
and utility methods.
 
 
 
-Connection
-Admin.getConnection()
-
-
 (package private) Connection
 RegionAdminServiceCallable.getConnection()
 
-
+
 protected Connection
 HTable.getConnection()
 INTERNAL Used by unit tests and tools to do 
low-level
  manipulations.
 
 
+
+Connection
+Admin.getConnection()
+
 
 Connection
 HBaseAdmin.getConnection()
@@ -1506,11 +1506,11 @@ Input/OutputFormats, a table indexing MapReduce job, 
and utility methods.
 
 
 private Connection
-TableInputFormatBase.connection
+HRegionPartitioner.connection
 
 
 private Connection
-HRegionPartitioner.connection
+TableInputFormatBase.connection
 
 
 
@@ -1543,22 +1543,22 @@ Input/OutputFormats, a table indexing MapReduce job, 
and utility methods.
 
 
 
-private Connection
-TableOutputFormat.TableRecordWriter.connection
-
-
 (package private) Connection
 MultiTableOutputFormat.MultiTableRecordWriter.connection
 
+
+private Connection
+HRegionPartitioner.connection
+
 
 private Connection
-TableInputFormatBase.connection
-The underlying Connection 
of the table.
-
+TableOutputFormat.TableRecordWriter.connection
 
 
 private Connection
-HRegionPartitioner.connection
+TableInputFormatBase.connection
+The underlying Connection 
of the table.
+
 
 
 (package private) Connection
@@ -1643,15 +1643,15 @@ Input/OutputFormats, a table indexing MapReduce job, 
and utility methods.
 
 
 private Connection
-CatalogJanitor.connection
+RegionPlacementMaintainer.connection
 
 
 private Connection
-SnapshotOfRegionAssignmentFromMeta.connection
+CatalogJanitor.connection
 
 
 private Connection
-RegionPlacementMaintainer.connection
+SnapshotOfRegionAssignmentFromMeta.connection
 
 
 
@@ -1787,16 +1787,16 @@ Input/OutputFormats, a table indexing MapReduce job, 
and utility methods.
 
 
 
-private Connection
-TableQuotaSnapshotStore.conn
+(package private) Connection
+FileArchiverNotifierFactoryImpl.CacheKey.conn
 
 
 private Connection
-SpaceQuotaRefresherChore.conn
+QuotaObserverChore.conn
 
 
 private Connection
-NamespaceQuotaSnapshotStore.conn
+QuotaObserverChore.TablesWithQuotas.conn
 
 
 private Connection
@@ -1804,11 +1804,11 @@ Input/OutputFormats, a table indexing MapReduce job, 
and utility methods.
 
 
 private Connection
-QuotaObserverChore.conn
+NamespaceQuotaSnapshotStore.conn
 
 
 private Connection
-QuotaObserverChore.TablesWithQuotas.conn
+TableQuotaSnapshotStore.conn
 
 
 private Connection
@@ -1816,6 +1816,14 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 private Connection
+FileArchiverNotifierImpl.conn
+
+
+private Connection
+SpaceQuotaRefresherChore.conn
+
+
+private Connection
 QuotaRetriever.connection
 Connection to use.
 
@@ -1972,6 +1980,24 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListGetgets)
 
 
+FileArchiverNotifier
+FileArchiverNotifierFactoryImpl.get(Connectionconn,
+   org.apache.hadoop.conf.Configurationconf,
+   org.apache.hadoop.fs.FileSystemfs,
+   TableNametn)
+Returns the FileArchiverNotifier instance 
for the given TableName.
+
+
+
+FileArchiverNotifier
+FileArchiverNotifierFactory.get(Connectionconn,
+   org.apache.hadoop.conf.Configurationconf,
+   org.apache.hadoop.fs.FileSystemfs,
+   TableNametn)
+Creates or obtains a FileArchiverNotifier instance 
for the given args.
+
+
+
 static SpaceQuotaSnapshot
 QuotaTableUtil.getCurrentSnapshot(Connectionconn,
   

[23/51] [partial] hbase-site git commit: Published site at b7b86839250bf9b295ebc1948826f43a88736d6c.

2018-03-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html 
b/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html
index 42307d2..2ade137 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9,"i40":6,"i41":9,"i42":9,"i43":9,"i44":9,"i45":9,"i46":9,"i47":9,"i48":9};
+var methods = 
{"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9,"i40":9,"i41":6,"i42":9,"i43":9,"i44":9,"i45":9,"i46":9,"i47":9,"i48":9,"i49":9};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],4:["t3","Abstract 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -429,6 +429,11 @@ extends 
 
 
+static org.apache.hadoop.fs.Path
+getRegionDir(org.apache.hadoop.fs.PathtableDir,
+RegionInforegion)
+
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.Path
 getRegionDirs(org.apache.hadoop.fs.FileSystemfs,
  org.apache.hadoop.fs.PathtableDir)
@@ -436,7 +441,7 @@ extends 
 
 
-
+
 private static void
 getRegionLocalityMappingFromFS(org.apache.hadoop.conf.Configurationconf,
   https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringdesiredTable,
@@ -449,17 +454,17 @@ extends 
 
 
-
+
 static int
 getRegionReferenceFileCount(org.apache.hadoop.fs.FileSystemfs,
org.apache.hadoop.fs.Pathp)
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.Path
 getTableDirs(org.apache.hadoop.fs.FileSystemfs,
 org.apache.hadoop.fs.Pathrootdir)
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer
 getTableFragmentation(org.apache.hadoop.fs.FileSystemfs,
  org.apache.hadoop.fs.PathhbaseRootDir)
@@ -467,14 +472,14 @@ extends 
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer
 getTableFragmentation(HMastermaster)
 Runs through the HBase rootdir and checks how many stores 
for each table
  have more than one file in them.
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.fs.Path
 getTableStoreFilePathMap(org.apache.hadoop.fs.FileSystemfs,
 org.apache.hadoop.fs.PathhbaseRootDir)
@@ -482,7 +487,7 @@ extends 
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.fs.Path
 getTableStoreFilePathMap(org.apache.hadoop.fs.FileSystemfs,
 org.apache.hadoop.fs.PathhbaseRootDir,
@@ -493,7 +498,7 @@ extends 
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.fs.Path
 

[23/51] [partial] hbase-site git commit: Published site at 1d25b60831b8cc8f7ad5fd366f1867de5c20d2f3.

2018-03-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html
--
diff --git 
a/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html
 
b/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html
index e12d889..afdcd68 100644
--- 
a/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html
+++ 
b/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html
@@ -134,7 +134,7 @@ public static interface Method and Description
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalCursor
+https://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalCursor
 cursor()
 Get the scan cursor if available.
 
@@ -202,7 +202,7 @@ public static interface 
 
 cursor
-http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalCursorcursor()
+https://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalCursorcursor()
 Get the scan cursor if available.
 
 Returns:

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/client/Append.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/client/Append.html 
b/apidocs/org/apache/hadoop/hbase/client/Append.html
index e41805f..b655827 100644
--- a/apidocs/org/apache/hadoop/hbase/client/Append.html
+++ b/apidocs/org/apache/hadoop/hbase/client/Append.html
@@ -97,7 +97,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
 
 
 org.apache.hadoop.hbase.client.Operation
@@ -124,7 +124,7 @@ var activeTableTab = "activeTableTab";
 
 
 All Implemented Interfaces:
-http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableRow, CellScannable, Attributes, Row, 
org.apache.hadoop.hbase.io.HeapSize
+https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableRow, CellScannable, Attributes, Row, 
org.apache.hadoop.hbase.io.HeapSize
 
 
 
@@ -206,7 +206,7 @@ extends 
 Append(byte[]row,
   longts,
-  http://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true;
 title="class or interface in java.util">NavigableMapbyte[],http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListCellfamilyMap)
+  https://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true;
 title="class or interface in java.util">NavigableMapbyte[],https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListCellfamilyMap)
 Construct the Append with user defined data.
 
 
@@ -268,16 +268,16 @@ extends 
 
 Append
-setACL(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.Permissionperms)
+setACL(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">String,org.apache.hadoop.hbase.security.access.Permissionperms)
 
 
 Append
-setACL(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringuser,
+setACL(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringuser,
   
org.apache.hadoop.hbase.security.access.Permissionperms)
 
 
 Append
-setAttribute(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringname,
+setAttribute(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringname,
 byte[]value)
 Sets an attribute.
 
@@ -290,7 +290,7 @@ extends 
 
 Append

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html
 
b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html
index 6cba39a..3a47680 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html
@@ -103,7 +103,7 @@ var activeTableTab = "activeTableTab";
 http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true;
 title="class or interface in java.lang">java.lang.Thread
 
 
-org.apache.hadoop.hbase.procedure2.ProcedureExecutor.StoppableThread
+org.apache.hadoop.hbase.procedure2.StoppableThread
 
 
 org.apache.hadoop.hbase.procedure2.ProcedureExecutor.WorkerThread
@@ -122,13 +122,17 @@ var activeTableTab = "activeTableTab";
 http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true;
 title="class or interface in java.lang">Runnable
 
 
+Direct Known Subclasses:
+ProcedureExecutor.KeepAliveWorkerThread
+
+
 Enclosing class:
 ProcedureExecutorTEnvironment
 
 
 
-private final class ProcedureExecutor.WorkerThread
-extends ProcedureExecutor.StoppableThread
+private class ProcedureExecutor.WorkerThread
+extends StoppableThread
 
 
 
@@ -163,7 +167,7 @@ extends Field and Description
 
 
-private Procedure
+private Procedure?
 activeProcedure
 
 
@@ -189,10 +193,17 @@ extends 
 Constructors
 
-Constructor and Description
+Modifier
+Constructor and Description
 
 
-WorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true;
 title="class or interface in 
java.lang">ThreadGroupgroup)
+
+WorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true;
 title="class or interface in 
java.lang">ThreadGroupgroup)
+
+
+protected 
+WorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true;
 title="class or interface in java.lang">ThreadGroupgroup,
+http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringprefix)
 
 
 
@@ -214,7 +225,7 @@ extends getCurrentRunTime()
 
 
-private boolean
+protected boolean
 keepAlive(longlastUpdate)
 
 
@@ -231,11 +242,11 @@ extends 
-
+
 
 
-Methods inherited from classorg.apache.hadoop.hbase.procedure2.ProcedureExecutor.StoppableThread
-awaitTermination
+Methods inherited from classorg.apache.hadoop.hbase.procedure2.StoppableThread
+awaitTermination
 
 
 
@@ -271,7 +282,7 @@ extends 
 
 executionStartTime
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true;
 title="class or interface in java.util.concurrent.atomic">AtomicLong executionStartTime
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true;
 title="class or interface in java.util.concurrent.atomic">AtomicLong executionStartTime
 
 
 
@@ -280,7 +291,7 @@ extends 
 
 activeProcedure
-privateProcedure activeProcedure
+private volatileProcedure? activeProcedure
 
 
 
@@ -294,10 +305,20 @@ extends 
 
 
+
+
+WorkerThread
+publicWorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true;
 title="class or interface in java.lang">ThreadGroupgroup)
+
+
+
+
+
 
 
 WorkerThread
-publicWorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true;
 title="class or interface in java.lang">ThreadGroupgroup)
+protectedWorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true;
 title="class or interface in java.lang">ThreadGroupgroup,
+   http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringprefix)
 
 
 
@@ -314,10 +335,10 @@ extends 
 
 sendStopSignal
-publicvoidsendStopSignal()
+publicvoidsendStopSignal()
 
 Specified by:
-sendStopSignalin
 classProcedureExecutor.StoppableThread
+sendStopSignalin
 classStoppableThread
 
 
 
@@ -327,7 +348,7 @@ extends 
 
 run
-publicvoidrun()
+publicvoidrun()
 
 Specified by:
 http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--;
 title="class or interface in java.lang">runin 
interfacehttp://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true;
 title="class or interface in java.lang">Runnable
@@ -342,7 +363,7 @@ extends 
 
 toString
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtoString()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or 

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
index 70c0a84..f458c4c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
@@ -263,541 +263,547 @@
 255throw new 
UnsupportedOperationException(this + " unhandled state=" + state);
 256  }
 257} catch (IOException e) {
-258  LOG.warn("Error trying to merge 
regions " + RegionInfo.getShortNameToLog(regionsToMerge) +
-259" in the table " + getTableName() 
+ " (in state=" + state + ")", e);
-260
-261  setFailure("master-merge-regions", 
e);
-262}
-263return Flow.HAS_MORE_STATE;
-264  }
-265
-266  /**
-267   * To rollback {@link 
MergeTableRegionsProcedure}, two AssignProcedures are asynchronously
-268   * submitted for each region to be 
merged (rollback doesn't wait on the completion of the
-269   * AssignProcedures) . This can be 
improved by changing rollback() to support sub-procedures.
-270   * See HBASE-19851 for details.
-271   */
-272  @Override
-273  protected void rollbackState(
-274  final MasterProcedureEnv env,
-275  final MergeTableRegionsState state) 
throws IOException, InterruptedException {
-276if (isTraceEnabled()) {
-277  LOG.trace(this + " rollback state=" 
+ state);
-278}
-279
-280try {
-281  switch (state) {
-282  case 
MERGE_TABLE_REGIONS_POST_OPERATION:
-283  case 
MERGE_TABLE_REGIONS_OPEN_MERGED_REGION:
-284  case 
MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION:
-285  case 
MERGE_TABLE_REGIONS_UPDATE_META:
-286String msg = this + " We are in 
the " + state + " state."
-287+ " It is complicated to 
rollback the merge operation that region server is working on."
-288+ " Rollback is not supported 
and we should let the merge operation to complete";
-289LOG.warn(msg);
-290// PONR
-291throw new 
UnsupportedOperationException(this + " unhandled state=" + state);
-292  case 
MERGE_TABLE_REGIONS_PRE_MERGE_COMMIT_OPERATION:
-293break;
-294  case 
MERGE_TABLE_REGIONS_CREATE_MERGED_REGION:
-295cleanupMergedRegion(env);
-296break;
-297  case 
MERGE_TABLE_REGIONS_CLOSE_REGIONS:
-298
rollbackCloseRegionsForMerge(env);
+258  String msg = "Error trying to merge 
regions " +
+259
RegionInfo.getShortNameToLog(regionsToMerge) + " in the table " + 
getTableName() +
+260   " (in state=" + state + ")";
+261  if (!isRollbackSupported(state)) 
{
+262// We reach a state that cannot 
be rolled back. We just need to keep retry.
+263LOG.warn(msg, e);
+264  } else {
+265LOG.error(msg, e);
+266
setFailure("master-merge-regions", e);
+267  }
+268}
+269return Flow.HAS_MORE_STATE;
+270  }
+271
+272  /**
+273   * To rollback {@link 
MergeTableRegionsProcedure}, two AssignProcedures are asynchronously
+274   * submitted for each region to be 
merged (rollback doesn't wait on the completion of the
+275   * AssignProcedures) . This can be 
improved by changing rollback() to support sub-procedures.
+276   * See HBASE-19851 for details.
+277   */
+278  @Override
+279  protected void rollbackState(
+280  final MasterProcedureEnv env,
+281  final MergeTableRegionsState state) 
throws IOException, InterruptedException {
+282if (isTraceEnabled()) {
+283  LOG.trace(this + " rollback state=" 
+ state);
+284}
+285
+286try {
+287  switch (state) {
+288  case 
MERGE_TABLE_REGIONS_POST_OPERATION:
+289  case 
MERGE_TABLE_REGIONS_OPEN_MERGED_REGION:
+290  case 
MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION:
+291  case 
MERGE_TABLE_REGIONS_UPDATE_META:
+292String msg = this + " We are in 
the " + state + " state."
+293+ " It is complicated to 
rollback the merge operation that region server is working on."
+294+ " Rollback is not supported 
and we should let the merge operation to complete";
+295LOG.warn(msg);
+296// PONR
+297throw new 
UnsupportedOperationException(this + " unhandled state=" + state);
+298  case 
MERGE_TABLE_REGIONS_PRE_MERGE_COMMIT_OPERATION:
 299break;
-300  case 
MERGE_TABLE_REGIONS_PRE_MERGE_OPERATION:
-301postRollBackMergeRegions(env);
+300  case 
MERGE_TABLE_REGIONS_CREATE_MERGED_REGION:
+301cleanupMergedRegion(env);
 302break;
-303  case MERGE_TABLE_REGIONS_PREPARE:
-304break;
-305  default:
-306throw new 

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-17 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
index 6d8c565..7edb3ff 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
@@ -2726,863 +2726,873 @@
 2718}
 2719  }
 2720
-2721  void checkInitialized()
-2722  throws PleaseHoldException, 
ServerNotRunningYetException, MasterNotRunningException {
-2723checkServiceStarted();
-2724if (!isInitialized()) throw new 
PleaseHoldException("Master is initializing");
-2725if (isStopped()) throw new 
MasterNotRunningException();
-2726  }
-2727
-2728  /**
-2729   * Report whether this master is 
currently the active master or not.
-2730   * If not active master, we are parked 
on ZK waiting to become active.
-2731   *
-2732   * This method is used for testing.
-2733   *
-2734   * @return true if active master, 
false if not.
-2735   */
-2736  @Override
-2737  public boolean isActiveMaster() {
-2738return activeMaster;
-2739  }
-2740
-2741  /**
-2742   * Report whether this master has 
completed with its initialization and is
-2743   * ready.  If ready, the master is 
also the active master.  A standby master
-2744   * is never ready.
-2745   *
-2746   * This method is used for testing.
-2747   *
-2748   * @return true if master is ready to 
go, false if not.
-2749   */
-2750  @Override
-2751  public boolean isInitialized() {
-2752return initialized.isReady();
-2753  }
-2754
-2755  /**
-2756   * Report whether this master is in 
maintenance mode.
+2721  public static class 
MasterStoppedException extends DoNotRetryIOException {
+2722MasterStoppedException() {
+2723  super();
+2724}
+2725  }
+2726
+2727  void checkInitialized() throws 
PleaseHoldException, ServerNotRunningYetException,
+2728  MasterNotRunningException, 
MasterStoppedException {
+2729checkServiceStarted();
+2730if (!isInitialized()) {
+2731  throw new 
PleaseHoldException("Master is initializing");
+2732}
+2733if (isStopped()) {
+2734  throw new 
MasterStoppedException();
+2735}
+2736  }
+2737
+2738  /**
+2739   * Report whether this master is 
currently the active master or not.
+2740   * If not active master, we are parked 
on ZK waiting to become active.
+2741   *
+2742   * This method is used for testing.
+2743   *
+2744   * @return true if active master, 
false if not.
+2745   */
+2746  @Override
+2747  public boolean isActiveMaster() {
+2748return activeMaster;
+2749  }
+2750
+2751  /**
+2752   * Report whether this master has 
completed with its initialization and is
+2753   * ready.  If ready, the master is 
also the active master.  A standby master
+2754   * is never ready.
+2755   *
+2756   * This method is used for testing.
 2757   *
-2758   * @return true if master is in 
maintenanceMode
+2758   * @return true if master is ready to 
go, false if not.
 2759   */
 2760  @Override
-2761  public boolean isInMaintenanceMode() 
{
-2762return 
maintenanceModeTracker.isInMaintenanceMode();
+2761  public boolean isInitialized() {
+2762return initialized.isReady();
 2763  }
 2764
-2765  @VisibleForTesting
-2766  public void setInitialized(boolean 
isInitialized) {
-2767
procedureExecutor.getEnvironment().setEventReady(initialized, isInitialized);
-2768  }
-2769
+2765  /**
+2766   * Report whether this master is in 
maintenance mode.
+2767   *
+2768   * @return true if master is in 
maintenanceMode
+2769   */
 2770  @Override
-2771  public ProcedureEvent? 
getInitializedEvent() {
-2772return initialized;
+2771  public boolean isInMaintenanceMode() 
{
+2772return 
maintenanceModeTracker.isInMaintenanceMode();
 2773  }
 2774
-2775  /**
-2776   * ServerCrashProcessingEnabled is set 
false before completing assignMeta to prevent processing
-2777   * of crashed servers.
-2778   * @return true if assignMeta has 
completed;
-2779   */
+2775  @VisibleForTesting
+2776  public void setInitialized(boolean 
isInitialized) {
+2777
procedureExecutor.getEnvironment().setEventReady(initialized, isInitialized);
+2778  }
+2779
 2780  @Override
-2781  public boolean 
isServerCrashProcessingEnabled() {
-2782return 
serverCrashProcessingEnabled.isReady();
+2781  public ProcedureEvent? 
getInitializedEvent() {
+2782return initialized;
 2783  }
 2784
-2785  @VisibleForTesting
-2786  public void 
setServerCrashProcessingEnabled(final boolean b) {
-2787
procedureExecutor.getEnvironment().setEventReady(serverCrashProcessingEnabled, 
b);
-2788  }
-2789
-2790  public ProcedureEvent? 
getServerCrashProcessingEnabledEvent() {
-2791return 

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html
index 4584cda..fb9bdb3 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html
@@ -137,9 +137,7 @@
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
-AsyncAdmin.listSnapshots()
-List completed snapshots.
-
+AsyncHBaseAdmin.listSnapshots()
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
@@ -148,22 +146,22 @@
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
-RawAsyncHBaseAdmin.listSnapshots()
-
-
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
 HBaseAdmin.listSnapshots()
 
+
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
+AsyncAdmin.listSnapshots()
+List completed snapshots.
+
+
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
-AsyncHBaseAdmin.listSnapshots()
+RawAsyncHBaseAdmin.listSnapshots()
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
-AsyncAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true;
 title="class or interface in java.util.regex">Patternpattern)
-List all the completed snapshots matching the given 
pattern.
-
+AsyncHBaseAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true;
 title="class or interface in 
java.util.regex">Patternpattern)
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
@@ -172,16 +170,18 @@
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
-RawAsyncHBaseAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true;
 title="class or interface in 
java.util.regex">Patternpattern)
-
-
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
 HBaseAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true;
 title="class or interface in 
java.util.regex">Patternpattern)
 
+
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListSnapshotDescription
+AsyncAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true;
 title="class or interface in java.util.regex">Patternpattern)
+List all the completed snapshots matching the given 
pattern.
+
+
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in 

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
index 232a8b4..b0d9cb7 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
@@ -495,7 +495,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 private static HRegionLocation
-AsyncMetaTableAccessor.getRegionLocation(Resultr,
+MetaTableAccessor.getRegionLocation(Resultr,
  RegionInforegionInfo,
  intreplicaId)
 Returns the HRegionLocation parsed from the given meta row 
Result
@@ -504,7 +504,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 private static HRegionLocation
-MetaTableAccessor.getRegionLocation(Resultr,
+AsyncMetaTableAccessor.getRegionLocation(Resultr,
  RegionInforegionInfo,
  intreplicaId)
 Returns the HRegionLocation parsed from the given meta row 
Result
@@ -944,7 +944,9 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-AsyncHBaseAdmin.getRegions(ServerNameserverName)
+AsyncAdmin.getRegions(ServerNameserverName)
+Get all the online regions on a region server.
+
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
@@ -953,22 +955,22 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-HBaseAdmin.getRegions(ServerNamesn)
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
+RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-AsyncAdmin.getRegions(ServerNameserverName)
-Get all the online regions on a region server.
-
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
+HBaseAdmin.getRegions(ServerNamesn)
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
+AsyncHBaseAdmin.getRegions(ServerNameserverName)
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-AsyncHBaseAdmin.getRegions(TableNametableName)
+AsyncAdmin.getRegions(TableNametableName)
+Get the regions of a given table.
+
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
@@ -977,18 +979,16 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-HBaseAdmin.getRegions(TableNametableName)
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
+RawAsyncHBaseAdmin.getRegions(TableNametableName)
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in 

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html
index 35f0e35..e3d9f70 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html
@@ -208,9 +208,9 @@ service.
 
 
 
-ResultScanner
-HTable.getScanner(byte[]family)
-The underlying HTable must 
not be closed.
+default ResultScanner
+AsyncTable.getScanner(byte[]family)
+Gets a scanner on the current table for the given 
family.
 
 
 
@@ -220,16 +220,16 @@ service.
 
 
 
-default ResultScanner
-AsyncTable.getScanner(byte[]family)
-Gets a scanner on the current table for the given 
family.
+ResultScanner
+HTable.getScanner(byte[]family)
+The underlying HTable must 
not be closed.
 
 
 
-ResultScanner
-HTable.getScanner(byte[]family,
+default ResultScanner
+AsyncTable.getScanner(byte[]family,
   byte[]qualifier)
-The underlying HTable must 
not be closed.
+Gets a scanner on the current table for the given family 
and qualifier.
 
 
 
@@ -240,37 +240,37 @@ service.
 
 
 
-default ResultScanner
-AsyncTable.getScanner(byte[]family,
+ResultScanner
+HTable.getScanner(byte[]family,
   byte[]qualifier)
-Gets a scanner on the current table for the given family 
and qualifier.
+The underlying HTable must 
not be closed.
 
 
 
 ResultScanner
-RawAsyncTableImpl.getScanner(Scanscan)
-
-
-ResultScanner
-HTable.getScanner(Scanscan)
-The underlying HTable must 
not be closed.
+AsyncTable.getScanner(Scanscan)
+Returns a scanner on the current table as specified by the 
Scan 
object.
 
 
-
+
 ResultScanner
 Table.getScanner(Scanscan)
 Returns a scanner on the current table as specified by the 
Scan
  object.
 
 
-
+
 ResultScanner
 AsyncTableImpl.getScanner(Scanscan)
 
+
+ResultScanner
+RawAsyncTableImpl.getScanner(Scanscan)
+
 
 ResultScanner
-AsyncTable.getScanner(Scanscan)
-Returns a scanner on the current table as specified by the 
Scan 
object.
+HTable.getScanner(Scanscan)
+The underlying HTable must 
not be closed.
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
index d730879..b1d1cef 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
@@ -106,11 +106,11 @@
 
 
 RetriesExhaustedWithDetailsException
-AsyncRequestFutureImpl.getErrors()
+AsyncRequestFuture.getErrors()
 
 
 RetriesExhaustedWithDetailsException
-AsyncRequestFuture.getErrors()
+AsyncRequestFutureImpl.getErrors()
 
 
 (package private) RetriesExhaustedWithDetailsException

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html
index 9642faa..0a290e1 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html
@@ -234,36 +234,28 @@
 
 
 
-T
-RpcRetryingCallerImpl.callWithoutRetries(RetryingCallableTcallable,
-  intcallTimeout)
-
-
 T
 RpcRetryingCaller.callWithoutRetries(RetryingCallableTcallable,
   intcallTimeout)
 Call the server once only.
 
 
-
+
 T
-RpcRetryingCallerImpl.callWithRetries(RetryingCallableTcallable,
-   intcallTimeout)
+RpcRetryingCallerImpl.callWithoutRetries(RetryingCallableTcallable,
+  intcallTimeout)
 
-
+
 T
 RpcRetryingCaller.callWithRetries(RetryingCallableTcallable,
intcallTimeout)
 Retries if invocation fails.
 
 
-
-RetryingCallerInterceptorContext
-NoOpRetryingInterceptorContext.prepare(RetryingCallable?callable)
-
 
-FastFailInterceptorContext
-FastFailInterceptorContext.prepare(RetryingCallable?callable)
+T
+RpcRetryingCallerImpl.callWithRetries(RetryingCallableTcallable,
+   intcallTimeout)
 
 
 abstract RetryingCallerInterceptorContext
@@ -275,13 +267,11 @@
 
 
 RetryingCallerInterceptorContext
-NoOpRetryingInterceptorContext.prepare(RetryingCallable?callable,
-   inttries)

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-13 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html
index 7676dbd..dd70c74 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html
@@ -399,10 +399,8 @@ service.
 
 
 
-default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
-AsyncTable.exists(Getget)
-Test for the existence of columns in the table, as 
specified by the Get.
-
+boolean
+HTable.exists(Getget)
 
 
 boolean
@@ -411,32 +409,34 @@ service.
 
 
 
-boolean
-HTable.exists(Getget)
+default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
+AsyncTable.exists(Getget)
+Test for the existence of columns in the table, as 
specified by the Get.
+
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureResult
-AsyncTable.get(Getget)
-Extracts certain cells from a given row.
-
+RawAsyncTableImpl.get(Getget)
 
 
 Result
+HTable.get(Getget)
+
+
+Result
 Table.get(Getget)
 Extracts certain cells from a given row.
 
 
-
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureResult
-AsyncTableImpl.get(Getget)
-
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureResult
-RawAsyncTableImpl.get(Getget)
+AsyncTableImpl.get(Getget)
 
 
-Result
-HTable.get(Getget)
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureResult
+AsyncTable.get(Getget)
+Extracts certain cells from a given row.
+
 
 
 private Result
@@ -457,10 +457,8 @@ service.
 
 
 
-default http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
-AsyncTable.exists(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListGetgets)
-Test for the existence of columns in the table, as 
specified by the Gets.
-
+boolean[]
+HTable.exists(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListGetgets)
 
 
 boolean[]
@@ -469,16 +467,12 @@ service.
 
 
 
-boolean[]
-HTable.exists(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListGetgets)
-
-
-default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
-AsyncTable.existsAll(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListGetgets)
-A simple version for batch exists.
+default http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
+AsyncTable.exists(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListGetgets)
+Test for the existence of columns in the table, as 
specified by the Gets.
 
 
-

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html
index 35f0e35..e3d9f70 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html
@@ -208,9 +208,9 @@ service.
 
 
 
-ResultScanner
-HTable.getScanner(byte[]family)
-The underlying HTable must 
not be closed.
+default ResultScanner
+AsyncTable.getScanner(byte[]family)
+Gets a scanner on the current table for the given 
family.
 
 
 
@@ -220,16 +220,16 @@ service.
 
 
 
-default ResultScanner
-AsyncTable.getScanner(byte[]family)
-Gets a scanner on the current table for the given 
family.
+ResultScanner
+HTable.getScanner(byte[]family)
+The underlying HTable must 
not be closed.
 
 
 
-ResultScanner
-HTable.getScanner(byte[]family,
+default ResultScanner
+AsyncTable.getScanner(byte[]family,
   byte[]qualifier)
-The underlying HTable must 
not be closed.
+Gets a scanner on the current table for the given family 
and qualifier.
 
 
 
@@ -240,37 +240,37 @@ service.
 
 
 
-default ResultScanner
-AsyncTable.getScanner(byte[]family,
+ResultScanner
+HTable.getScanner(byte[]family,
   byte[]qualifier)
-Gets a scanner on the current table for the given family 
and qualifier.
+The underlying HTable must 
not be closed.
 
 
 
 ResultScanner
-RawAsyncTableImpl.getScanner(Scanscan)
-
-
-ResultScanner
-HTable.getScanner(Scanscan)
-The underlying HTable must 
not be closed.
+AsyncTable.getScanner(Scanscan)
+Returns a scanner on the current table as specified by the 
Scan 
object.
 
 
-
+
 ResultScanner
 Table.getScanner(Scanscan)
 Returns a scanner on the current table as specified by the 
Scan
  object.
 
 
-
+
 ResultScanner
 AsyncTableImpl.getScanner(Scanscan)
 
+
+ResultScanner
+RawAsyncTableImpl.getScanner(Scanscan)
+
 
 ResultScanner
-AsyncTable.getScanner(Scanscan)
-Returns a scanner on the current table as specified by the 
Scan 
object.
+HTable.getScanner(Scanscan)
+The underlying HTable must 
not be closed.
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
index d730879..b1d1cef 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html
@@ -106,11 +106,11 @@
 
 
 RetriesExhaustedWithDetailsException
-AsyncRequestFutureImpl.getErrors()
+AsyncRequestFuture.getErrors()
 
 
 RetriesExhaustedWithDetailsException
-AsyncRequestFuture.getErrors()
+AsyncRequestFutureImpl.getErrors()
 
 
 (package private) RetriesExhaustedWithDetailsException

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html
index 9642faa..0a290e1 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html
@@ -234,36 +234,28 @@
 
 
 
-T
-RpcRetryingCallerImpl.callWithoutRetries(RetryingCallableTcallable,
-  intcallTimeout)
-
-
 T
 RpcRetryingCaller.callWithoutRetries(RetryingCallableTcallable,
   intcallTimeout)
 Call the server once only.
 
 
-
+
 T
-RpcRetryingCallerImpl.callWithRetries(RetryingCallableTcallable,
-   intcallTimeout)
+RpcRetryingCallerImpl.callWithoutRetries(RetryingCallableTcallable,
+  intcallTimeout)
 
-
+
 T
 RpcRetryingCaller.callWithRetries(RetryingCallableTcallable,
intcallTimeout)
 Retries if invocation fails.
 
 
-
-RetryingCallerInterceptorContext
-NoOpRetryingInterceptorContext.prepare(RetryingCallable?callable)
-
 
-FastFailInterceptorContext
-FastFailInterceptorContext.prepare(RetryingCallable?callable)
+T
+RpcRetryingCallerImpl.callWithRetries(RetryingCallableTcallable,
+   intcallTimeout)
 
 
 abstract RetryingCallerInterceptorContext
@@ -275,13 +267,11 @@
 
 
 RetryingCallerInterceptorContext
-NoOpRetryingInterceptorContext.prepare(RetryingCallable?callable,
-   inttries)

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html
index d481372..5e1590b 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html
@@ -126,15 +126,15 @@
 
 
 private RpcRetryingCallerFactory
-RegionCoprocessorRpcChannel.rpcCallerFactory
+ConnectionImplementation.rpcCallerFactory
 
 
 private RpcRetryingCallerFactory
-ConnectionImplementation.rpcCallerFactory
+HTable.rpcCallerFactory
 
 
 private RpcRetryingCallerFactory
-HTable.rpcCallerFactory
+RegionCoprocessorRpcChannel.rpcCallerFactory
 
 
 private RpcRetryingCallerFactory
@@ -155,21 +155,21 @@
 
 
 RpcRetryingCallerFactory
-ConnectionImplementation.getNewRpcRetryingCallerFactory(org.apache.hadoop.conf.Configurationconf)
-
-
-RpcRetryingCallerFactory
 ClusterConnection.getNewRpcRetryingCallerFactory(org.apache.hadoop.conf.Configurationconf)
 Returns a new RpcRetryingCallerFactory from the given 
Configuration.
 
 
+
+RpcRetryingCallerFactory
+ConnectionImplementation.getNewRpcRetryingCallerFactory(org.apache.hadoop.conf.Configurationconf)
+
 
 RpcRetryingCallerFactory
-ConnectionImplementation.getRpcRetryingCallerFactory()
+ClusterConnection.getRpcRetryingCallerFactory()
 
 
 RpcRetryingCallerFactory
-ClusterConnection.getRpcRetryingCallerFactory()
+ConnectionImplementation.getRpcRetryingCallerFactory()
 
 
 static RpcRetryingCallerFactory

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html
index 6384833..018438c 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html
@@ -283,27 +283,27 @@ service.
 
 
 private Scan
-ScannerCallableWithReplicas.scan
+AsyncScanSingleRegionRpcRetryingCaller.scan
 
 
 protected Scan
-ClientScanner.scan
+ScannerCallable.scan
 
 
 private Scan
-AsyncClientScanner.scan
+ScannerCallableWithReplicas.scan
 
 
-private Scan
-AsyncRpcRetryingCallerFactory.ScanSingleRegionCallerBuilder.scan
+protected Scan
+ClientScanner.scan
 
 
 private Scan
-AsyncScanSingleRegionRpcRetryingCaller.scan
+AsyncClientScanner.scan
 
 
-protected Scan
-ScannerCallable.scan
+private Scan
+AsyncRpcRetryingCallerFactory.ScanSingleRegionCallerBuilder.scan
 
 
 private Scan
@@ -339,11 +339,11 @@ service.
 
 
 protected Scan
-ClientScanner.getScan()
+ScannerCallable.getScan()
 
 
 protected Scan
-ScannerCallable.getScan()
+ClientScanner.getScan()
 
 
 Scan
@@ -638,29 +638,29 @@ service.
 
 
 ResultScanner
-AsyncTable.getScanner(Scanscan)
-Returns a scanner on the current table as specified by the 
Scan 
object.
-
+RawAsyncTableImpl.getScanner(Scanscan)
 
 
 ResultScanner
-Table.getScanner(Scanscan)
-Returns a scanner on the current table as specified by the 
Scan
- object.
+HTable.getScanner(Scanscan)
+The underlying HTable must 
not be closed.
 
 
 
 ResultScanner
-AsyncTableImpl.getScanner(Scanscan)
+Table.getScanner(Scanscan)
+Returns a scanner on the current table as specified by the 
Scan
+ object.
+
 
 
 ResultScanner
-RawAsyncTableImpl.getScanner(Scanscan)
+AsyncTableImpl.getScanner(Scanscan)
 
 
 ResultScanner
-HTable.getScanner(Scanscan)
-The underlying HTable must 
not be closed.
+AsyncTable.getScanner(Scanscan)
+Returns a scanner on the current table as specified by the 
Scan 
object.
 
 
 
@@ -703,9 +703,7 @@ service.
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListResult
-AsyncTable.scanAll(Scanscan)
-Return all the results that match the given scan 
object.
-
+RawAsyncTableImpl.scanAll(Scanscan)
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListResult
@@ -713,7 +711,9 @@ service.
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in 

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html
index ad601c4..53e455f 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html
@@ -1117,1183 +1117,1186 @@
 1109  @Nullable
 1110  public static TableState 
getTableState(Connection conn, TableName tableName)
   throws IOException {
-1112Table metaHTable = 
getMetaHTable(conn);
-1113Get get = new 
Get(tableName.getName()).addColumn(getTableFamily(), getTableStateColumn());
-1114long time = 
EnvironmentEdgeManager.currentTime();
-1115get.setTimeRange(0, time);
-1116Result result =
-1117metaHTable.get(get);
-1118return getTableState(result);
-1119  }
-1120
-1121  /**
-1122   * Fetch table states from META 
table
-1123   * @param conn connection to use
-1124   * @return map {tableName -gt; 
state}
-1125   * @throws IOException
-1126   */
-1127  public static MapTableName, 
TableState getTableStates(Connection conn)
-1128  throws IOException {
-1129final MapTableName, 
TableState states = new LinkedHashMap();
-1130Visitor collector = new Visitor() 
{
-1131  @Override
-1132  public boolean visit(Result r) 
throws IOException {
-1133TableState state = 
getTableState(r);
-1134if (state != null)
-1135  
states.put(state.getTableName(), state);
-1136return true;
-1137  }
-1138};
-1139fullScanTables(conn, collector);
-1140return states;
-1141  }
-1142
-1143  /**
-1144   * Updates state in META
-1145   * @param conn connection to use
-1146   * @param tableName table to look 
for
-1147   * @throws IOException
-1148   */
-1149  public static void 
updateTableState(Connection conn, TableName tableName,
-1150  TableState.State actual) throws 
IOException {
-1151updateTableState(conn, new 
TableState(tableName, actual));
-1152  }
-1153
-1154  /**
-1155   * Decode table state from META 
Result.
-1156   * Should contain cell from 
HConstants.TABLE_FAMILY
-1157   * @param r result
-1158   * @return null if not found
-1159   * @throws IOException
-1160   */
-1161  @Nullable
-1162  public static TableState 
getTableState(Result r)
-1163  throws IOException {
-1164Cell cell = 
r.getColumnLatestCell(getTableFamily(), getTableStateColumn());
-1165if (cell == null) return null;
-1166try {
-1167  return 
TableState.parseFrom(TableName.valueOf(r.getRow()),
-1168  
Arrays.copyOfRange(cell.getValueArray(),
-1169  cell.getValueOffset(), 
cell.getValueOffset() + cell.getValueLength()));
-1170} catch (DeserializationException e) 
{
-1171  throw new IOException(e);
-1172}
-1173
-1174  }
-1175
-1176  /**
-1177   * Implementations 'visit' a catalog 
table row.
-1178   */
-1179  public interface Visitor {
-1180/**
-1181 * Visit the catalog table row.
-1182 * @param r A row from catalog 
table
-1183 * @return True if we are to proceed 
scanning the table, else false if
-1184 * we are to stop now.
-1185 */
-1186boolean visit(final Result r) throws 
IOException;
-1187  }
-1188
-1189  /**
-1190   * Implementations 'visit' a catalog 
table row but with close() at the end.
-1191   */
-1192  public interface CloseableVisitor 
extends Visitor, Closeable {
-1193  }
-1194
-1195  /**
-1196   * A {@link Visitor} that collects 
content out of passed {@link Result}.
-1197   */
-1198  static abstract class 
CollectingVisitorT implements Visitor {
-1199final ListT results = new 
ArrayList();
-1200@Override
-1201public boolean visit(Result r) 
throws IOException {
-1202  if (r ==  null || r.isEmpty()) 
return true;
-1203  add(r);
-1204  return true;
-1205}
-1206
-1207abstract void add(Result r);
-1208
-1209/**
-1210 * @return Collected results; wait 
till visits complete to collect all
-1211 * possible results
-1212 */
-1213ListT getResults() {
-1214  return this.results;
-1215}
-1216  }
-1217
-1218  /**
-1219   * Collects all returned.
-1220   */
-1221  static class CollectAllVisitor extends 
CollectingVisitorResult {
-1222@Override
-1223void add(Result r) {
-1224  this.results.add(r);
-1225}
-1226  }
-1227
-1228  /**
-1229   * A Visitor that skips offline 
regions and split parents
-1230   */
-1231  public static abstract class 
DefaultVisitorBase implements Visitor {
-1232
-1233public DefaultVisitorBase() {
-1234  super();
-1235}
-1236
-1237public abstract boolean 
visitInternal(Result rowResult) throws IOException;
-1238
-1239@Override
-1240public boolean visit(Result 
rowResult) throws IOException {
-1241  RegionInfo info 

[23/51] [partial] hbase-site git commit: Published site at .

2018-02-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
index ca0e581..7eb86ac 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
@@ -535,14 +535,14 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
+org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer
 (implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
 org.apache.hadoop.hbase.util.ChecksumType
-org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer
 (implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
 org.apache.hadoop.hbase.util.PrettyPrinter.Unit
-org.apache.hadoop.hbase.util.Order
-org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer
 (implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
 org.apache.hadoop.hbase.util.PoolMap.PoolType
 org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
 org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE
+org.apache.hadoop.hbase.util.Order
+org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer
 (implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html 
b/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html
index c08dcaa..e1d8b82 100644
--- a/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html
+++ b/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":6,"i2":6,"i3":9,"i4":9,"i5":9,"i6":10,"i7":9,"i8":10,"i9":9,"i10":9,"i11":9,"i12":10,"i13":9,"i14":9,"i15":9,"i16":10,"i17":10,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":10,"i25":9};
+var methods = 
{"i0":10,"i1":10,"i2":6,"i3":6,"i4":9,"i5":9,"i6":9,"i7":10,"i8":9,"i9":10,"i10":9,"i11":9,"i12":9,"i13":10,"i14":9,"i15":9,"i16":9,"i17":10,"i18":10,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":10,"i26":9};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],4:["t3","Abstract 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -283,71 +283,77 @@ implements 
 
 void
+addWALActionsListener(WALActionsListenerlistener)
+Add a WALActionsListener.
+
+
+
+void
 close()
 shutdown utstanding WALs and clean up any persisted 
state.
 
 
-
+
 protected abstract T
 createWAL()
 
-
+
 protected abstract void
 doInit(org.apache.hadoop.conf.Configurationconf)
 
-
+
 static long
 extractFileNumFromWAL(WALwal)
 It returns the file create timestamp from the file 
name.
 
 
-
+
 static org.apache.hadoop.fs.Path
 getArchivedLogPath(org.apache.hadoop.fs.Pathpath,
   org.apache.hadoop.conf.Configurationconf)
 Get the archived WAL file path
 
 
-
+
 static org.apache.hadoop.fs.Path
 getCurrentFileName(WALwal)
 return the current filename from the current wal.
 
 
-
+
 long
 getLogFileSize()
 iff the given WALFactory is using the DefaultWALProvider 
for meta and/or non-meta, count the
  size of files (only rolled).
 
 
-
+
 static long
 getLogFileSize(WALwal)
 returns the size of rolled WAL files.
 
 
-
+
 long
 getNumLogFiles()
 iff the given WALFactory is using the DefaultWALProvider 
for meta and/or non-meta, count the
  number of files (rolled and active).
 
 
-
+
 static int
 getNumRolledLogFiles(WALwal)
 returns the number of rolled WAL files.
 
 
-
+
 static ServerName
 getServerNameFromWALDirectoryName(org.apache.hadoop.conf.Configurationconf,
  http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringpath)
 Pulls a ServerName out of a Path generated according to our 
layout rules.
 
 
-
+
 static ServerName
 getServerNameFromWALDirectoryName(org.apache.hadoop.fs.PathlogFile)
 This function returns region server name from a log file 
name which is in one of the following
@@ -358,81 +364,79 @@ implements 
 
 
-
+
 T
 getWAL(RegionInforegion)
 
-
+
 static 

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html
 
b/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html
index 46c2cc7..416914b 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public class TestRepairAfterFailedDelete
+public class TestRepairAfterFailedDelete
 extends TestBackupBase
 
 
@@ -150,6 +150,10 @@ extends Field and Description
 
 
+static HBaseClassTestRule
+CLASS_RULE
+
+
 private static org.slf4j.Logger
 LOG
 
@@ -225,13 +229,22 @@ extends 
+
+
+
+
+CLASS_RULE
+public static finalHBaseClassTestRule CLASS_RULE
+
+
 
 
 
 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
@@ -248,7 +261,7 @@ extends 
 
 TestRepairAfterFailedDelete
-publicTestRepairAfterFailedDelete()
+publicTestRepairAfterFailedDelete()
 
 
 
@@ -265,7 +278,7 @@ extends 
 
 testRepairBackupDelete
-publicvoidtestRepairBackupDelete()
+publicvoidtestRepairBackupDelete()
 throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exception
 
 Throws:

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html 
b/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html
index 0297e64..1267c07 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html
@@ -150,6 +150,10 @@ extends Field and Description
 
 
+static HBaseClassTestRule
+CLASS_RULE
+
+
 private static org.slf4j.Logger
 LOG
 
@@ -233,13 +237,22 @@ extends 
+
+
+
+
+CLASS_RULE
+public static finalHBaseClassTestRule CLASS_RULE
+
+
 
 
 
 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
@@ -273,7 +286,7 @@ extends 
 
 testFullRestoreSingleEmpty
-publicvoidtestFullRestoreSingleEmpty()
+publicvoidtestFullRestoreSingleEmpty()
 throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exception
 Verify that a single empty table is restored to a new 
table.
 
@@ -288,7 +301,7 @@ extends 
 
 testFullRestoreMultipleEmpty
-publicvoidtestFullRestoreMultipleEmpty()
+publicvoidtestFullRestoreMultipleEmpty()
   throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exception
 Verify that multiple tables are restored to new 
tables.
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html 
b/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html
index 4d45c8f..ef25995 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public class TestSystemTableSnapshot
+public class TestSystemTableSnapshot
 extends TestBackupBase
 
 
@@ -150,6 +150,10 @@ extends Field and Description
 
 
+static HBaseClassTestRule
+CLASS_RULE
+
+
 private static org.slf4j.Logger
 LOG
 
@@ -227,13 +231,22 @@ extends 
+
+
+
+
+CLASS_RULE
+public static finalHBaseClassTestRule CLASS_RULE
+
+
 
 
 
 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
@@ -250,7 +263,7 @@ extends 
 
 TestSystemTableSnapshot
-publicTestSystemTableSnapshot()
+publicTestSystemTableSnapshot()
 
 
 
@@ -267,7 +280,7 @@ extends 
 
 _testBackupRestoreSystemTable
-publicvoid_testBackupRestoreSystemTable()
+publicvoid_testBackupRestoreSystemTable()
throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exception
 Verify backup system table snapshot.
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.html

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html
index eb9e252..667152a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html
@@ -28,22 +28,22 @@
 020
 021import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
 022import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
-023import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
-024import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
-025import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
-026import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
-027import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
-028import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
-029import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
-030import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
-031import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
-032import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
-033import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
-034import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
-035import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
-036import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
-037import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
-038import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
+023import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
+024import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
+025import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
+026import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
+027import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
+028import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
+029import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
+030import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
+031import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
+032import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
+033import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
+034import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
+035import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
+036import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
+037import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
+038import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
 039
 040import java.io.IOException;
 041import java.net.URI;
@@ -70,194 +70,194 @@
 062import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
 063import 
org.apache.hadoop.hbase.client.Connection;
 064import 
org.apache.hadoop.hbase.client.ConnectionFactory;
-065import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
-066import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-067import 
org.apache.yetus.audience.InterfaceAudience;
-068
-069/**
-070 * General backup commands, options and 
usage messages
-071 */
-072
+065import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+066import 
org.apache.yetus.audience.InterfaceAudience;
+067
+068import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
+069
+070/**
+071 * General backup commands, options and 
usage messages
+072 */
 073@InterfaceAudience.Private
 074public final class BackupCommands {
-075
-076  public final static String 
INCORRECT_USAGE = "Incorrect usage";
-077
-078  public final static String 
TOP_LEVEL_NOT_ALLOWED =
-079  "Top level (root) folder is not 
allowed to be a backup destination";
-080
-081  public static final String USAGE = 
"Usage: hbase backup 

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html
index 914b1c6..03a0b2a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html
@@ -59,646 +59,711 @@
 051@InterfaceStability.Evolving
 052public class ScannerContext {
 053
-054  /**
-055   * Two sets of the same fields. One for 
the limits, another for the progress towards those limits
-056   */
-057  LimitFields limits;
-058  LimitFields progress;
-059
-060  /**
-061   * The state of the scanner after the 
invocation of {@link InternalScanner#next(java.util.List)}
-062   * or {@link 
RegionScanner#next(java.util.List)}.
-063   */
-064  NextState scannerState;
-065  private static final NextState 
DEFAULT_STATE = NextState.MORE_VALUES;
-066
-067  /**
-068   * Used as an indication to invocations 
of {@link InternalScanner#next(java.util.List)} and
-069   * {@link 
RegionScanner#next(java.util.List)} that, if true, the progress tracked within 
this
-070   * {@link ScannerContext} instance 
should be considered while evaluating the limits. Useful for
-071   * enforcing a set of limits across 
multiple calls (i.e. the limit may not be reached in a single
-072   * invocation, but any progress made 
should be considered in future invocations)
-073   * p
-074   * Defaulting this value to false means 
that, by default, any tracked progress will be wiped clean
-075   * on invocations to {@link 
InternalScanner#next(java.util.List)} and
-076   * {@link 
RegionScanner#next(java.util.List)} and the call will be treated as though no 
progress
-077   * has been made towards the limits so 
far.
-078   * p
-079   * This is an important mechanism. 
Users of Internal/Region scanners expect that they can define
-080   * some limits and then repeatedly 
invoke {@link InternalScanner#next(List)} or
-081   * {@link RegionScanner#next(List)} 
where each invocation respects these limits separately.
-082   * p
-083   * For example: pre {@code
-084   * ScannerContext context = new 
ScannerContext.newBuilder().setBatchLimit(5).build();
-085   * RegionScanner scanner = ...
-086   * ListCell results = new 
ArrayListCell();
-087   * while(scanner.next(results, 
context)) {
-088   *   // Do something with a batch of 5 
cells
-089   * }
-090   * }/pre However, in the case 
of RPCs, the server wants to be able to define a set of
-091   * limits for a particular RPC request 
and have those limits respected across multiple
-092   * invocations. This means that the 
progress made towards the limits in earlier calls will be
-093   * saved and considered in future 
invocations
-094   */
-095  boolean keepProgress;
-096  private static boolean 
DEFAULT_KEEP_PROGRESS = false;
-097
-098  private Cell lastPeekedCell = null;
+054  LimitFields limits;
+055  /**
+056   * A different set of progress fields. 
Only include batch, dataSize and heapSize. Compare to
+057   * LimitFields, ProgressFields doesn't 
contain time field. As we save a deadline in LimitFields,
+058   * so use {@link 
System#currentTimeMillis()} directly when check time limit.
+059   */
+060  ProgressFields progress;
+061
+062  /**
+063   * The state of the scanner after the 
invocation of {@link InternalScanner#next(java.util.List)}
+064   * or {@link 
RegionScanner#next(java.util.List)}.
+065   */
+066  NextState scannerState;
+067  private static final NextState 
DEFAULT_STATE = NextState.MORE_VALUES;
+068
+069  /**
+070   * Used as an indication to invocations 
of {@link InternalScanner#next(java.util.List)} and
+071   * {@link 
RegionScanner#next(java.util.List)} that, if true, the progress tracked within 
this
+072   * {@link ScannerContext} instance 
should be considered while evaluating the limits. Useful for
+073   * enforcing a set of limits across 
multiple calls (i.e. the limit may not be reached in a single
+074   * invocation, but any progress made 
should be considered in future invocations)
+075   * p
+076   * Defaulting this value to false means 
that, by default, any tracked progress will be wiped clean
+077   * on invocations to {@link 
InternalScanner#next(java.util.List)} and
+078   * {@link 
RegionScanner#next(java.util.List)} and the call will be treated as though no 
progress
+079   * has been made towards the limits so 
far.
+080   * p
+081   * This is an important mechanism. 
Users of Internal/Region scanners expect that they can define
+082   * some limits and then repeatedly 
invoke {@link InternalScanner#next(List)} or
+083   * {@link RegionScanner#next(List)} 
where each invocation respects these limits 

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html 
b/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html
index d880bf7..3d34dcf 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html
@@ -332,7 +332,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 ServerNameDeathDateComparator
-private statichttp://docs.oracle.com/javase/8/docs/api/java/util/Comparator.html?is-external=true;
 title="class or interface in java.util">ComparatorPairServerName,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long ServerNameDeathDateComparator
+private statichttp://docs.oracle.com/javase/8/docs/api/java/util/Comparator.html?is-external=true;
 title="class or interface in java.util">ComparatorPairServerName,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long ServerNameDeathDateComparator
 
 
 
@@ -488,7 +488,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 toString
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtoString()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtoString()
 
 Overrides:
 http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
 title="class or interface in java.lang">toStringin 
classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
@@ -501,7 +501,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 copyDeadServersSince
-publichttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListPairServerName,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">LongcopyDeadServersSince(longts)
+publichttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListPairServerName,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">LongcopyDeadServersSince(longts)
 Extract all the servers dead since a given time, and sort 
them.
 
 Parameters:
@@ -517,7 +517,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getTimeOfDeath
-publichttp://docs.oracle.com/javase/8/docs/api/java/util/Date.html?is-external=true;
 title="class or interface in java.util">DategetTimeOfDeath(ServerNamedeadServerName)
+publichttp://docs.oracle.com/javase/8/docs/api/java/util/Date.html?is-external=true;
 title="class or interface in java.util">DategetTimeOfDeath(ServerNamedeadServerName)
 Get the time when a server died
 
 Parameters:
@@ -533,7 +533,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 removeDeadServer
-publicbooleanremoveDeadServer(ServerNamedeadServerName)
+publicbooleanremoveDeadServer(ServerNamedeadServerName)
 remove the specified dead server
 
 Parameters:



[23/51] [partial] hbase-site git commit: Published site at .

2018-01-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmplImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmplImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmplImpl.html
new file mode 100644
index 000..1f26c87
--- /dev/null
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmplImpl.html
@@ -0,0 +1,560 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001// Autogenerated Jamon 
implementation
+002// 
/home/jenkins/jenkins-slave/workspace/hbase_generate_website/hbase/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmpl.jamon
+003
+004package 
org.apache.hadoop.hbase.tmpl.master;
+005
+006// 27, 5
+007import java.util.Collections;
+008// 28, 5
+009import java.util.List;
+010// 29, 5
+011import java.util.Map;
+012// 30, 5
+013import java.util.Set;
+014// 31, 5
+015import java.util.stream.Collectors;
+016// 32, 5
+017import 
org.apache.hadoop.hbase.master.HMaster;
+018// 33, 5
+019import 
org.apache.hadoop.hbase.ServerLoad;
+020// 34, 5
+021import 
org.apache.hadoop.hbase.RSGroupTableAccessor;
+022// 35, 5
+023import 
org.apache.hadoop.hbase.master.ServerManager;
+024// 36, 5
+025import 
org.apache.hadoop.hbase.net.Address;
+026// 37, 5
+027import 
org.apache.hadoop.hbase.rsgroup.RSGroupInfo;
+028// 38, 5
+029import 
org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix;
+030
+031public class RSGroupListTmplImpl
+032  extends 
org.jamon.AbstractTemplateImpl
+033  implements 
org.apache.hadoop.hbase.tmpl.master.RSGroupListTmpl.Intf
+034
+035{
+036  private final HMaster master;
+037  private final ServerManager 
serverManager;
+038  protected static 
org.apache.hadoop.hbase.tmpl.master.RSGroupListTmpl.ImplData 
__jamon_setOptionalArguments(org.apache.hadoop.hbase.tmpl.master.RSGroupListTmpl.ImplData
 p_implData)
+039  {
+040return p_implData;
+041  }
+042  public 
RSGroupListTmplImpl(org.jamon.TemplateManager p_templateManager, 
org.apache.hadoop.hbase.tmpl.master.RSGroupListTmpl.ImplData p_implData)
+043  {
+044super(p_templateManager, 
__jamon_setOptionalArguments(p_implData));
+045master = p_implData.getMaster();
+046serverManager = 
p_implData.getServerManager();
+047  }
+048  
+049  @Override public void 
renderNoFlush(final java.io.Writer jamonWriter)
+050throws java.io.IOException
+051  {
+052// 40, 1
+053
+054ListRSGroupInfo groups = 
RSGroupTableAccessor.getAllRSGroupInfo(master.getConnection());
+055
+056// 44, 1
+057if ((groups != null  
groups.size()  0))
+058{
+059  // 44, 45
+060  jamonWriter.write("\n\n");
+061  // 46, 1
+062  
+063RSGroupInfo [] rsGroupInfos = 
groups.toArray(new RSGroupInfo[groups.size()]);
+064MapAddress, ServerLoad 
collectServers = Collections.emptyMap();
+065if (master.getServerManager() != null) 
{
+066  collectServers =
+067  
master.getServerManager().getOnlineServers().entrySet().stream()
+068  .collect(Collectors.toMap(p 
- p.getKey().getAddress(), Map.Entry::getValue));
+069}
+070
+071  // 56, 1
+072  jamonWriter.write("div 
class=\"tabbable\"\nul class=\"nav nav-pills\"\nli 
class=\"active\"a href=\"#tab_rsgroup_baseStats\" 
data-toggle=\"tab\"Base Stats/a/li\nli 
class=\"\"a href=\"#tab_rsgroup_memoryStats\" 
data-toggle=\"tab\"Memory/a/li\nli 
class=\"\"a href=\"#tab_rsgroup_requestStats\" 
data-toggle=\"tab\"Requests/a/li\nli 
class=\"\"a href=\"#tab_rsgroup_storeStats\" 
data-toggle=\"tab\"Storefiles/a/li\nli 
class=\"\"a href=\"#tab_rsgroup_compactStats\" 
data-toggle=\"tab\"Compactions/a/li\n/ul\n
div class=\"tab-content\" style=\"padding-bottom: 9px; border-bottom: 1px 
solid #ddd;\"\ndiv class=\"tab-pane active\" 
id=\"tab_rsgroup_baseStats\"\n");
+073  // 66, 13
+074  {
+075// 66, 13
+076
__jamon_innerUnit__rsgroup_baseStats(jamonWriter, rsGroupInfos, collectServers 
);
+077  }
+078  // 66, 97
+079  jamonWriter.write("\n
/div\ndiv class=\"tab-pane\" 
id=\"tab_rsgroup_memoryStats\"\n");
+080  // 69, 13
+081  {
+082// 69, 13
+083
__jamon_innerUnit__rsgroup_memoryStats(jamonWriter, rsGroupInfos, 
collectServers );
+084  }
+085  // 69, 99
+086  jamonWriter.write("\n
/div\ndiv class=\"tab-pane\" 
id=\"tab_rsgroup_requestStats\"\n");
+087  // 72, 13
+088  {
+089// 72, 13
+090
__jamon_innerUnit__rsgroup_requestStats(jamonWriter, rsGroupInfos, 
collectServers );
+091  }
+092  // 72, 100
+093  jamonWriter.write("\n
/div\ndiv class=\"tab-pane\" 
id=\"tab_rsgroup_storeStats\"\n");
+094  // 75, 13
+095  {
+096// 75, 13

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html
 
b/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html
index 0048578..d5b9f50 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html
@@ -180,7 +180,7 @@ extends 
 
 Methods inherited from 
classorg.apache.hadoop.hbase.regionserver.HRegionServer
-abort, addRegion, addToMovedRegions, canCreateBaseZNode, 
canUpdateTableDescriptor, checkFileSystem, clearRegionBlockCache, 
closeAllRegions, closeAndOfflineRegionForSplitOrMerge, closeRegion, 
configureInfoServer, constructRegionServer, convertThrowableToIOE, 
createClusterConnection, createConnection, createRegionLoad, 
createRegionServerStatusStub, createRegionServerStatusStub, createRpcServices, 
execRegionServerService, executeProcedure, getCacheConfig, getChoreService, 
getClusterConnection, getClusterId, getCompactionPressure, 
getCompactionRequestor, getCompactSplitThread, getConfiguration, 
getConfigurationManager, getConnection, getCoordinatedStateManager, 
getDumpServlet, getEventLoopGroupConfig, getExecutorService, 
getFavoredNodesForRegion, getFileSystem, getFlushPressure, getFlushRequester, 
getFlushThroughputController, getFsTableDescriptors, getHeapMemoryManager, 
getInfoServer, getLastSequenceId, getLeases, getMasterAddressTracker, 
getMetaTableLocator, getMetaTableObserver
 , getMetrics, getMostLoadedRegions, getNonceManager, getNumberOfOnlineRegions, 
getOnlineRegion, getOnlineRegionsLocalContext, getOnlineTables, getProcessName, 
getRegion, getRegion, getRegionBlockLocations, getRegionByEncodedName, 
getRegionByEncodedName, getRegions, getRegions, getRegionServerAccounting, 
getRegionServerCoprocessorHost, getRegionServerCoprocessors, 
getRegionServerMetrics, getRegionServerRpcQuotaManager, 
getRegionServerSpaceQuotaManager, getRegionsInTransitionInRS, 
getReplicationSourceService, getRootDir, getRpcServer, getRSRpcServices, 
getSecureBulkLoadManager, getServerName, getStartcode, getTableDescriptors, 
getThreadWakeFrequency, getWAL, getWALFileSystem, getWalRoller, getWALRootDir, 
getWALs, getZooKeeper, initializeMemStoreChunkCreator, isAborted, isOnline, 
isStopped, isStopping, login, main, onConfigurationChange, postOpenDeployTasks, 
regionLock, registerService, remoteProcedureComplete, removeRegion, 
reportRegionSizesForQuotas, reportRegionStateTransition, send
 ShutdownInterrupt, setInitLatch, setupClusterConnection, 
shouldUseThisHostnameInstead, stop, stop, stopServiceThreads, toString, 
tryRegionServerReport, unassign, updateConfiguration, 
updateRegionFavoredNodesMapping, waitForMasterActive, waitForServerOnline, 
walRollRequestFinished
+abort, addRegion, addToMovedRegions, canCreateBaseZNode, 
canUpdateTableDescriptor, checkFileSystem, clearRegionBlockCache, 
closeAllRegions, closeAndOfflineRegionForSplitOrMerge, closeRegion, 
configureInfoServer, constructRegionServer, convertThrowableToIOE, 
createClusterConnection, createConnection, createRegionLoad, 
createRegionServerStatusStub, createRegionServerStatusStub, createRpcServices, 
execRegionServerService, executeProcedure, getCacheConfig, getChoreService, 
getClusterConnection, getClusterId, getCompactionPressure, 
getCompactionRequestor, getCompactSplitThread, getConfiguration, 
getConfigurationManager, getConnection, getCoordinatedStateManager, 
getDumpServlet, getEventLoopGroupConfig, getExecutorService, 
getFavoredNodesForRegion, getFileSystem, getFlushPressure, getFlushRequester, 
getFlushThroughputController, getFsTableDescriptors, getHeapMemoryManager, 
getInfoServer, getLastSequenceId, getLeases, getMasterAddressTracker, 
getMetaTableLocator, getMetaTableObserver
 , getMetrics, getMostLoadedRegions, getNonceManager, getNumberOfOnlineRegions, 
getOnlineRegion, getOnlineRegionsLocalContext, getOnlineTables, getProcessName, 
getRegion, getRegion, getRegionBlockLocations, getRegionByEncodedName, 
getRegionByEncodedName, getRegions, getRegions, getRegionServerAccounting, 
getRegionServerCoprocessorHost, getRegionServerCoprocessors, 
getRegionServerMetrics, getRegionServerRpcQuotaManager, 
getRegionServerSpaceQuotaManager, getRegionsInTransitionInRS, 
getReplicationSourceService, getRootDir, getRpcServer, getRSRpcServices, 
getSecureBulkLoadManager, getServerName, getStartcode, getTableDescriptors, 
getThreadWakeFrequency, getWAL, getWALFileSystem, getWalRoller, getWALRootDir, 
getWALs, getZooKeeper, initializeMemStoreChunkCreator, isAborted, isOnline, 
isStopped, isStopping, login, main, onConfigurationChange, postOpenDeployTasks, 
regionLock, registerService, 

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-11 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
index fadf667..14b2b69 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
@@ -7,269 +7,269 @@
 
 
 001/**
-002 *
-003 * Licensed to the Apache Software 
Foundation (ASF) under one
-004 * or more contributor license 
agreements.  See the NOTICE file
-005 * distributed with this work for 
additional information
-006 * regarding copyright ownership.  The 
ASF licenses this file
-007 * to you under the Apache License, 
Version 2.0 (the
-008 * "License"); you may not use this file 
except in compliance
-009 * with the License.  You may obtain a 
copy of the License at
-010 *
-011 * 
http://www.apache.org/licenses/LICENSE-2.0
-012 *
-013 * Unless required by applicable law or 
agreed to in writing, software
-014 * distributed under the License is 
distributed on an "AS IS" BASIS,
-015 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
-016 * See the License for the specific 
language governing permissions and
-017 * limitations under the License.
-018 */
+002 * Licensed to the Apache Software 
Foundation (ASF) under one
+003 * or more contributor license 
agreements.  See the NOTICE file
+004 * distributed with this work for 
additional information
+005 * regarding copyright ownership.  The 
ASF licenses this file
+006 * to you under the Apache License, 
Version 2.0 (the
+007 * "License"); you may not use this file 
except in compliance
+008 * with the License.  You may obtain a 
copy of the License at
+009 *
+010 * 
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or 
agreed to in writing, software
+013 * distributed under the License is 
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+015 * See the License for the specific 
language governing permissions and
+016 * limitations under the License.
+017 */
+018package org.apache.hadoop.hbase.wal;
 019
-020
-021package org.apache.hadoop.hbase.wal;
-022
-023import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-024
-025import java.io.IOException;
-026import java.io.InterruptedIOException;
-027import java.util.Collections;
-028import java.util.List;
-029import java.util.OptionalLong;
-030import 
java.util.concurrent.atomic.AtomicReference;
-031
-032import 
org.apache.hadoop.conf.Configuration;
-033import org.apache.hadoop.fs.FileSystem;
-034import org.apache.hadoop.fs.Path;
-035import 
org.apache.yetus.audience.InterfaceAudience;
-036import org.slf4j.Logger;
-037import org.slf4j.LoggerFactory;
-038// imports for things that haven't moved 
from regionserver.wal yet.
-039import 
org.apache.hadoop.hbase.regionserver.wal.MetricsWAL;
-040import 
org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader;
-041import 
org.apache.hadoop.hbase.regionserver.wal.WALActionsListener;
-042import 
org.apache.hadoop.hbase.replication.regionserver.WALFileLengthProvider;
-043import 
org.apache.hadoop.hbase.util.CancelableProgressable;
-044import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-045import 
org.apache.hadoop.hbase.util.LeaseNotRecoveredException;
-046import 
org.apache.hadoop.hbase.wal.WAL.Reader;
-047import 
org.apache.hadoop.hbase.wal.WALProvider.Writer;
-048
-049/**
-050 * Entry point for users of the Write 
Ahead Log.
-051 * Acts as the shim between internal use 
and the particular WALProvider we use to handle wal
-052 * requests.
-053 *
-054 * Configure which provider gets used 
with the configuration setting "hbase.wal.provider". Available
-055 * implementations:
-056 * ul
-057 *   
liemdefaultProvider/em : whatever provider is standard 
for the hbase version. Currently
-058 *  
"filesystem"/li
-059 *   
liemfilesystem/em : a provider that will run on top of 
an implementation of the Hadoop
-060 * FileSystem 
interface, normally HDFS./li
-061 *   
liemmultiwal/em : a provider that will use multiple 
"filesystem" wal instances per region
-062 *   
server./li
-063 * /ul
-064 *
-065 * Alternatively, you may provide a 
custom implementation of {@link WALProvider} by class name.
-066 */
-067@InterfaceAudience.Private
-068public class WALFactory implements 
WALFileLengthProvider {
-069
-070  private static final Logger LOG = 
LoggerFactory.getLogger(WALFactory.class);
-071
-072  /**
-073   * Maps between configuration names for 
providers and implementation classes.
-074   */
-075  static enum Providers {
-076

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html
 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html
index 685363b..4a7ef76 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html
@@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-PrevClass
+PrevClass
 NextClass
 
 
@@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private class RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer
+private class RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer
 extends RawAsyncHBaseAdmin.TableProcedureBiConsumer
 
 
@@ -232,7 +232,7 @@ extends 
 
 SplitTableRegionProcedureBiConsumer
-SplitTableRegionProcedureBiConsumer(TableNametableName)
+SplitTableRegionProcedureBiConsumer(TableNametableName)
 
 
 
@@ -249,7 +249,7 @@ extends 
 
 getOperationType
-http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetOperationType()
+http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetOperationType()
 
 Specified by:
 getOperationTypein
 classRawAsyncHBaseAdmin.TableProcedureBiConsumer
@@ -284,7 +284,7 @@ extends 
 
-PrevClass
+PrevClass
 NextClass
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
index bd1c256..98a34ec 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
@@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
 
 
 http://docs.oracle.com/javase/8/docs/api/java/lang/FunctionalInterface.html?is-external=true;
 title="class or interface in java.lang">@FunctionalInterface
-private static interface RawAsyncHBaseAdmin.TableOperator
+private static interface RawAsyncHBaseAdmin.TableOperator
 
 
 
@@ -154,7 +154,7 @@ private static interface 
 
 operate
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Voidoperate(TableNametable)
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Voidoperate(TableNametable)
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html
 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html
index af56e24..d31719c 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html
@@ -126,7 +126,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private abstract class RawAsyncHBaseAdmin.TableProcedureBiConsumer
+private abstract class RawAsyncHBaseAdmin.TableProcedureBiConsumer
 extends RawAsyncHBaseAdmin.ProcedureBiConsumer
 
 
@@ -240,7 +240,7 @@ extends 
 
 tableName
-protected finalTableName tableName
+protected finalTableName tableName
 
 
 
@@ -257,7 +257,7 @@ extends 
 
 TableProcedureBiConsumer
-TableProcedureBiConsumer(TableNametableName)
+TableProcedureBiConsumer(TableNametableName)
 
 
 
@@ -274,7 +274,7 @@ extends 
 
 getOperationType
-abstracthttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetOperationType()
+abstracthttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetOperationType()
 
 
 
@@ -283,7 +283,7 @@ extends 
 
 getDescription

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
index 429ebaa..b4404ce 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
@@ -612,8 +612,8 @@
 
 
 default void
-MasterObserver.postGetClusterStatus(ObserverContextMasterCoprocessorEnvironmentctx,
-ClusterStatusstatus)
+MasterObserver.postGetClusterMetrics(ObserverContextMasterCoprocessorEnvironmentctx,
+ ClusterMetricsstatus)
 Called after get cluster status.
 
 
@@ -1452,7 +1452,7 @@
 
 
 default void
-MasterObserver.preGetClusterStatus(ObserverContextMasterCoprocessorEnvironmentctx)
+MasterObserver.preGetClusterMetrics(ObserverContextMasterCoprocessorEnvironmentctx)
 Called before get cluster status.
 
 
@@ -2678,166 +2678,162 @@
 
 
 void
-AccessController.preGetClusterStatus(ObserverContextMasterCoprocessorEnvironmentctx)
-
-
-void
 AccessController.preGetLocks(ObserverContextMasterCoprocessorEnvironmentctx)
 
-
+
 void
 AccessController.preGetNamespaceDescriptor(ObserverContextMasterCoprocessorEnvironmentctx,
  http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringnamespace)
 
-
+
 void
 AccessController.preGetOp(ObserverContextRegionCoprocessorEnvironmentc,
 Getget,
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListCellresult)
 
-
+
 void
 AccessController.preGetProcedures(ObserverContextMasterCoprocessorEnvironmentctx)
 
-
+
 void
 AccessController.preGetReplicationPeerConfig(ObserverContextMasterCoprocessorEnvironmentctx,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">StringpeerId)
 
-
+
 void
 AccessController.preGetTableDescriptors(ObserverContextMasterCoprocessorEnvironmentctx,
   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTableNametableNamesList,
   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTableDescriptordescriptors,
   http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringregex)
 
-
+
 Result
 AccessController.preIncrement(ObserverContextRegionCoprocessorEnvironmentc,
 Incrementincrement)
 
-
+
 Result
 AccessController.preIncrementAfterRowLock(ObserverContextRegionCoprocessorEnvironmentc,
 Incrementincrement)
 
-
+
 void
 AccessController.preListDecommissionedRegionServers(ObserverContextMasterCoprocessorEnvironmentctx)
 
-
+
 void
 AccessController.preListReplicationPeers(ObserverContextMasterCoprocessorEnvironmentctx,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringregex)
 
-
+
 void
 AccessController.preListSnapshot(ObserverContextMasterCoprocessorEnvironmentctx,
SnapshotDescriptionsnapshot)
 
-
+
 void
 AccessController.preLockHeartbeat(ObserverContextMasterCoprocessorEnvironmentctx,
 TableNametableName,
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringdescription)
 
-
+
 void
 AccessController.preMergeRegions(ObserverContextMasterCoprocessorEnvironmentctx,
RegionInfo[]regionsToMerge)
 
-
+
 void
 AccessController.preModifyNamespace(ObserverContextMasterCoprocessorEnvironmentctx,
   NamespaceDescriptorns)
 
-
+
 void
 AccessController.preModifyTable(ObserverContextMasterCoprocessorEnvironmentc,
   TableNametableName,
   TableDescriptorhtd)
 
-
+
 void
 CoprocessorWhitelistMasterObserver.preModifyTable(ObserverContextMasterCoprocessorEnvironmentctx,
   TableNametableName,
   TableDescriptorhtd)
 
-
+
 void
 AccessController.preMove(ObserverContextMasterCoprocessorEnvironmentc,
RegionInforegion,
ServerNamesrcServer,
ServerNamedestServer)
 
-
+
 void
 AccessController.preMoveServers(ObserverContextMasterCoprocessorEnvironmentctx,
   http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in java.util">SetAddressservers,
   

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
index 5b3b750..a1f3f7e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
@@ -97,3307 +97,3304 @@
 089import 
org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer;
 090import 
org.apache.hbase.thirdparty.io.netty.util.Timeout;
 091import 
org.apache.hbase.thirdparty.io.netty.util.TimerTask;
-092import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-093import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-094import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-095import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
-096import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
-098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse;
-099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
-105import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest;
-106import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse;
-107import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest;
-108import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse;
-109import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest;
-110import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse;
-111import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest;
-112import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse;
-113import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
-114import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
-115import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-116import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
-117import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
-118import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-119import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-120import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-121import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
-122import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
-123import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
-124import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest;
-125import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse;
-126import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-127import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-128import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest;
-129import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse;
-130import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest;
-131import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse;
-132import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
-133import 

[23/51] [partial] hbase-site git commit: Published site at .

2018-01-01 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html
index c8dbf3a..1116704 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html
@@ -248,6 +248,6 @@ public static interface Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html
index c89b6ca..721a4d2 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html
@@ -264,6 +264,6 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 
-Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html
 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html
index 280f7cf..5a38ff2 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html
@@ -649,6 +649,6 @@ extends Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html
 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html
index 65aa67d..c0888a4 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html
@@ -640,6 +640,6 @@ extends Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html
 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html
index 2a6aa03..9bc8eff 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html
@@ -640,6 +640,6 @@ extends Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html
index 07901b7..ff53947 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html
@@ -276,6 +276,6 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 
-Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-30 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html
index 985778f..854ba52 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html
@@ -662,1932 +662,1924 @@
 654
completeCompaction(toBeRemovedStoreFiles);
 655  }
 656
-657  private HStoreFile 
createStoreFileAndReader(final Path p) throws IOException {
-658StoreFileInfo info = new 
StoreFileInfo(conf, this.getFileSystem(), p);
-659return 
createStoreFileAndReader(info);
-660  }
-661
-662  private HStoreFile 
createStoreFileAndReader(StoreFileInfo info) throws IOException {
-663
info.setRegionCoprocessorHost(this.region.getCoprocessorHost());
-664HStoreFile storeFile = new 
HStoreFile(this.getFileSystem(), info, this.conf, this.cacheConf,
-665this.family.getBloomFilterType(), 
isPrimaryReplicaStore());
-666storeFile.initReader();
-667return storeFile;
-668  }
-669
-670  /**
-671   * This message intends to inform the 
MemStore that next coming updates
-672   * are going to be part of the 
replaying edits from WAL
-673   */
-674  public void startReplayingFromWAL(){
-675
this.memstore.startReplayingFromWAL();
-676  }
-677
-678  /**
-679   * This message intends to inform the 
MemStore that the replaying edits from WAL
-680   * are done
-681   */
-682  public void stopReplayingFromWAL(){
-683
this.memstore.stopReplayingFromWAL();
-684  }
-685
-686  /**
-687   * Adds a value to the memstore
-688   */
-689  public void add(final Cell cell, 
MemStoreSizing memstoreSizing) {
-690lock.readLock().lock();
-691try {
-692   this.memstore.add(cell, 
memstoreSizing);
-693} finally {
-694  lock.readLock().unlock();
-695}
-696  }
-697
-698  /**
-699   * Adds the specified value to the 
memstore
-700   */
-701  public void add(final 
IterableCell cells, MemStoreSizing memstoreSizing) {
-702lock.readLock().lock();
-703try {
-704  memstore.add(cells, 
memstoreSizing);
-705} finally {
-706  lock.readLock().unlock();
-707}
-708  }
-709
-710  @Override
-711  public long timeOfOldestEdit() {
-712return memstore.timeOfOldestEdit();
-713  }
-714
-715  /**
-716   * @return All store files.
-717   */
-718  @Override
-719  public CollectionHStoreFile 
getStorefiles() {
-720return 
this.storeEngine.getStoreFileManager().getStorefiles();
-721  }
-722
-723  @Override
-724  public CollectionHStoreFile 
getCompactedFiles() {
-725return 
this.storeEngine.getStoreFileManager().getCompactedfiles();
-726  }
-727
-728  /**
-729   * This throws a WrongRegionException 
if the HFile does not fit in this region, or an
-730   * InvalidHFileException if the HFile 
is not valid.
-731   */
-732  public void assertBulkLoadHFileOk(Path 
srcPath) throws IOException {
-733HFile.Reader reader  = null;
-734try {
-735  LOG.info("Validating hfile at " + 
srcPath + " for inclusion in "
-736  + "store " + this + " region " 
+ this.getRegionInfo().getRegionNameAsString());
-737  reader = 
HFile.createReader(srcPath.getFileSystem(conf), srcPath, cacheConf,
-738isPrimaryReplicaStore(), conf);
-739  reader.loadFileInfo();
-740
-741  Optionalbyte[] firstKey = 
reader.getFirstRowKey();
-742  
Preconditions.checkState(firstKey.isPresent(), "First key can not be null");
-743  OptionalCell lk = 
reader.getLastKey();
-744  
Preconditions.checkState(lk.isPresent(), "Last key can not be null");
-745  byte[] lastKey =  
CellUtil.cloneRow(lk.get());
-746
-747  LOG.debug("HFile bounds: first=" + 
Bytes.toStringBinary(firstKey.get()) +
-748  " last=" + 
Bytes.toStringBinary(lastKey));
-749  LOG.debug("Region bounds: first=" 
+
-750  
Bytes.toStringBinary(getRegionInfo().getStartKey()) +
-751  " last=" + 
Bytes.toStringBinary(getRegionInfo().getEndKey()));
-752
-753  if 
(!this.getRegionInfo().containsRange(firstKey.get(), lastKey)) {
-754throw new WrongRegionException(
-755"Bulk load file " + 
srcPath.toString() + " does not fit inside region "
-756+ 
this.getRegionInfo().getRegionNameAsString());
-757  }
-758
-759  if(reader.length()  
conf.getLong(HConstants.HREGION_MAX_FILESIZE,
-760  
HConstants.DEFAULT_MAX_FILE_SIZE)) {
-761LOG.warn("Trying to bulk load 
hfile " + srcPath.toString() + " with size: " +
-762reader.length() + " bytes can 
be problematic as it may lead to oversplitting.");
-763  }
-764
-765  if (verifyBulkLoads) {
-766long verificationStartTime = 
EnvironmentEdgeManager.currentTime();
-767LOG.info("Full verification 
started 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html
 
b/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html
index 7793de4..b03b542 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html
@@ -189,7 +189,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 Field and Description
 
 
-private static 
org.apache.hadoop.hbase.shaded.io.netty.buffer.ByteBufAllocator
+private static 
org.apache.hbase.thirdparty.io.netty.buffer.ByteBufAllocator
 ALLOC
 
 
@@ -293,8 +293,8 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 longfileId)
 
 
-private static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.hbase.shaded.io.netty.util.concurrent.Futureorg.apache.hadoop.hbase.shaded.io.netty.channel.Channel
-connectToDataNodes(org.apache.hadoop.conf.Configurationconf,
+private static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hbase.thirdparty.io.netty.util.concurrent.Futureorg.apache.hbase.thirdparty.io.netty.channel.Channel
+connectToDataNodes(org.apache.hadoop.conf.Configurationconf,
   org.apache.hadoop.hdfs.DFSClientclient,
   http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringclientName,
   
org.apache.hadoop.hdfs.protocol.LocatedBlocklocatedBlock,
@@ -302,8 +302,8 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
   longlatestGS,
   
org.apache.hadoop.hdfs.protocol.datatransfer.BlockConstructionStagestage,
   org.apache.hadoop.util.DataChecksumsummer,
-  
org.apache.hadoop.hbase.shaded.io.netty.channel.EventLoopGroupeventLoopGroup,
-  http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelchannelClass)
+  
org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroupeventLoopGroup,
+  http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? extends 
org.apache.hbase.thirdparty.io.netty.channel.ChannelchannelClass)
 
 
 private static FanOutOneBlockAsyncDFSOutputHelper.BlockAdder
@@ -349,27 +349,27 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 static FanOutOneBlockAsyncDFSOutput
-createOutput(org.apache.hadoop.hdfs.DistributedFileSystemdfs,
+createOutput(org.apache.hadoop.hdfs.DistributedFileSystemdfs,
 org.apache.hadoop.fs.Pathf,
 booleanoverwrite,
 booleancreateParent,
 shortreplication,
 longblockSize,
-
org.apache.hadoop.hbase.shaded.io.netty.channel.EventLoopGroupeventLoopGroup,
-http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelchannelClass)
+
org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroupeventLoopGroup,
+http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? extends 
org.apache.hbase.thirdparty.io.netty.channel.ChannelchannelClass)
 Create a FanOutOneBlockAsyncDFSOutput.
 
 
 
 private static FanOutOneBlockAsyncDFSOutput
-createOutput(org.apache.hadoop.hdfs.DistributedFileSystemdfs,
+createOutput(org.apache.hadoop.hdfs.DistributedFileSystemdfs,
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringsrc,
 booleanoverwrite,
 booleancreateParent,
 shortreplication,
 longblockSize,
-
org.apache.hadoop.hbase.shaded.io.netty.channel.EventLoopGroupeventLoopGroup,
-http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">Class? extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelchannelClass)
+
org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroupeventLoopGroup,
+http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/RawCell.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/RawCell.html 
b/devapidocs/org/apache/hadoop/hbase/RawCell.html
index a9b9734..0c030f3 100644
--- a/devapidocs/org/apache/hadoop/hbase/RawCell.html
+++ b/devapidocs/org/apache/hadoop/hbase/RawCell.html
@@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab";
 
 
 All Known Implementing Classes:
-BufferedDataBlockEncoder.OffheapDecodedCell,
 BufferedDataBlockEncoder.OnheapDecodedCell,
 ByteBufferChunkCell, 
ByteBufferKeyValue, IndividualBytesFieldCell, KeyValue, KeyValue.KeyOnlyKeyValue, MapReduceCell, Mutation.CellWrapper, NoTagByteBufferChunkCell, NoTagsByteBufferKeyValue, NoTagsKeyValue, Priv
 ateCellUtil.EmptyByteBufferCell, PrivateCellUtil.EmptyCell, PrivateCellUtil.FirstOnRowByteBufferCell, PrivateCellUtil.FirstOnRowCell, PrivateCellUtil.FirstOnRowColByteBufferCell, PrivateCellUtil.FirstOnRowColCell, PrivateCellUtil.FirstOnRowColTSByteBufferCell, PrivateCellUtil.FirstOnRowColTSCell, PrivateCellUtil.FirstOnRowDeleteFamilyCell, PrivateCellUtil.LastOnRowByteBufferCell, PrivateCellUtil.LastOnRowCell, PrivateCellUtil.LastOnRowColByteBufferCell, PrivateCellUtil.LastOnRowColCell, PrivateCellUtil.TagRewriteByteBufferCell, PrivateCellUtil.TagRewriteCell, PrivateCellUtil.ValueAndTagRewriteByteBufferCell, 
PrivateCellUtil.ValueAndTagRewriteCell, SizeCachedKeyValue, SizeCachedNoTagsKeyValue
+BufferedDataBlockEncoder.OffheapDecodedExtendedCell,
 BufferedDataBlockEncoder.OnheapDecodedCell,
 ByteBufferChunkKeyValue, ByteBufferExtendedCell, ByteBufferKeyOnlyKeyValue, ByteBufferKeyValue, IndividualBytesFieldCell, KeyOnlyFilter.KeyOnlyByteBufferExtendedCell,
 KeyValue, KeyValue.KeyOnlyKeyValue, MapReduceExtendedCell, Mutation.CellWrapper, NoTagByteBufferChunkKeyValue, NoTagsByteBufferKeyValue, NoTagsKeyValue, PrivateCellUtil.EmptyByteBufferExtendedCell, PrivateCellUtil.EmptyCell, PrivateCellUtil.FirstOnRowByteBufferExtendedCell, 
PrivateCellUtil.FirstOnRowCell, PrivateCellUtil.FirstOnRowColByteBufferExtendedCell,
 PrivateCellUtil.FirstOnRowColCell, PrivateCellUtil.FirstOnRowColTSByteBufferExtendedCell,
 PrivateCellUtil.FirstOnRowColTSCell, PrivateCellUtil.FirstOnR
 owDeleteFamilyCell, PrivateCellUtil.LastOnRowByteBufferExtendedCell, 
PrivateCellUtil.LastOnRowCell, PrivateCellUtil.LastOnRowColByteBufferExtendedCell,
 PrivateCellUtil.LastOnRowColCell, 
PrivateCellUtil.TagRewriteByteBufferExtendedCell, 
PrivateCellUtil.TagRewriteCell, PrivateCellUtil.ValueAndTagRewriteByteBufferExtendedCell,
 PrivateCellUtil.ValueAndTagRewriteCell, SizeCachedKeyValue, SizeCachedNoTagsKeyValue
 
 
 
@@ -134,7 +134,7 @@ extends Cell
-Cell.DataType
+Cell.Type
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html 
b/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html
index cc4b37f..d133996 100644
--- a/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html
+++ b/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html
@@ -185,7 +185,7 @@ extends 
 RawCellBuilder
-setType(Cell.DataTypetype)
+setType(Cell.Typetype)
 
 
 RawCellBuilder
@@ -309,16 +309,16 @@ extends 
+
 
 
 
 
 setType
-RawCellBuildersetType(Cell.DataTypetype)
+RawCellBuildersetType(Cell.Typetype)
 
 Specified by:
-setTypein
 interfaceCellBuilder
+setTypein
 interfaceCellBuilder
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html 
b/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html
index 3c48a21..3f4a0f7 100644
--- a/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html
+++ b/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html
@@ -123,11 +123,13 @@
 
 
 @InterfaceAudience.Public
-public class RegionTooBusyException
+public class RegionTooBusyException
 extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Thrown by a region server if it will block and wait to 
serve a request.
  For example, the client wants to insert something to a region while the
- region is compacting.
+ region is compacting. Keep variance in the passed 'msg' low because its msg 
is used as a key
+ over in RetriesExhaustedWithDetailsException
+ grouping failure types.
 
 See Also:
 Serialized
 Form
@@ -169,14 +171,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.ht
 Constructor and Description
 
 
-RegionTooBusyException()
-default constructor
-
-
-

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html 
b/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
index 9d0bd6f..afc9a5b 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
@@ -475,7 +475,7 @@ extends 
org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
 COMPRESSION_FAMILIES_CONF_KEY
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String COMPRESSION_FAMILIES_CONF_KEY
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String COMPRESSION_FAMILIES_CONF_KEY
 
 See Also:
 Constant
 Field Values
@@ -488,7 +488,7 @@ extends 
org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
 BLOOM_TYPE_FAMILIES_CONF_KEY
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String BLOOM_TYPE_FAMILIES_CONF_KEY
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String BLOOM_TYPE_FAMILIES_CONF_KEY
 
 See Also:
 Constant
 Field Values
@@ -501,7 +501,7 @@ extends 
org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
 BLOCK_SIZE_FAMILIES_CONF_KEY
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String BLOCK_SIZE_FAMILIES_CONF_KEY
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String BLOCK_SIZE_FAMILIES_CONF_KEY
 
 See Also:
 Constant
 Field Values
@@ -514,7 +514,7 @@ extends 
org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
 DATABLOCK_ENCODING_FAMILIES_CONF_KEY
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String DATABLOCK_ENCODING_FAMILIES_CONF_KEY
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String DATABLOCK_ENCODING_FAMILIES_CONF_KEY
 
 See Also:
 Constant
 Field Values
@@ -527,7 +527,7 @@ extends 
org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
 DATABLOCK_ENCODING_OVERRIDE_CONF_KEY
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String DATABLOCK_ENCODING_OVERRIDE_CONF_KEY
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String DATABLOCK_ENCODING_OVERRIDE_CONF_KEY
 
 See Also:
 Constant
 Field Values
@@ -540,7 +540,7 @@ extends 
org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
 LOCALITY_SENSITIVE_CONF_KEY
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String LOCALITY_SENSITIVE_CONF_KEY
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String LOCALITY_SENSITIVE_CONF_KEY
 Keep locality while generating HFiles for bulkload. See 
HBASE-12596
 
 See Also:
@@ -554,7 +554,7 @@ extends 
org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
 DEFAULT_LOCALITY_SENSITIVE
-private static finalboolean DEFAULT_LOCALITY_SENSITIVE
+private static finalboolean DEFAULT_LOCALITY_SENSITIVE
 
 See Also:
 Constant
 Field Values
@@ -567,7 +567,7 @@ extends 
org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
 OUTPUT_TABLE_NAME_CONF_KEY
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String OUTPUT_TABLE_NAME_CONF_KEY
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String OUTPUT_TABLE_NAME_CONF_KEY
 
 See Also:
 Constant
 Field Values
@@ -580,7 +580,7 @@ extends 
org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
 MULTI_TABLE_HFILEOUTPUTFORMAT_CONF_KEY
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MULTI_TABLE_HFILEOUTPUTFORMAT_CONF_KEY
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MULTI_TABLE_HFILEOUTPUTFORMAT_CONF_KEY
 
 See Also:
 Constant
 Field Values
@@ -593,7 +593,7 @@ extends 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html
 
b/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html
index 71630ed..7a5b5b6 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"]};
+var methods = 
{"i0":6,"i1":18,"i2":18,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -121,7 +121,7 @@ public interface 
-All MethodsInstance MethodsAbstract Methods
+All MethodsInstance MethodsAbstract MethodsDefault Methods
 
 Modifier and Type
 Method and Description
@@ -131,47 +131,57 @@ public interface build()
 
 
-ReplicationPeerConfigBuilder
-setBandwidth(longbandwidth)
+default ReplicationPeerConfigBuilder
+putAllConfiguration(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringconfiguration)
 
 
-ReplicationPeerConfigBuilder
-setClusterKey(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringclusterKey)
-Set the clusterKey which is the concatenation of the slave 
cluster's:
- 
hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
-
+default ReplicationPeerConfigBuilder
+putAllPeerData(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in 
java.util">Mapbyte[],byte[]peerData)
 
 
 ReplicationPeerConfigBuilder
-setConfiguration(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringconfiguration)
+putConfiguration(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringkey,
+http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringvalue)
 
 
 ReplicationPeerConfigBuilder
-setExcludeNamespaces(http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in java.util">Sethttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringnamespaces)
+putPeerData(byte[]key,
+   byte[]value)
 
 
 ReplicationPeerConfigBuilder
-setExcludeTableCFsMap(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">MapTableName,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">StringtableCFsMap)
+setBandwidth(longbandwidth)
 
 
 ReplicationPeerConfigBuilder
-setNamespaces(http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in java.util">Sethttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringnamespaces)
+setClusterKey(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringclusterKey)
+Set the clusterKey which is the concatenation of the slave 
cluster's:
+ 
hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
+
 
 
 ReplicationPeerConfigBuilder
-setPeerData(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in 
java.util">Mapbyte[],byte[]peerData)
+setExcludeNamespaces(http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
index 6fecbc9..2accda0 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
@@ -34,4140 +34,4141 @@
 026import 
java.nio.charset.StandardCharsets;
 027import java.util.ArrayList;
 028import java.util.Arrays;
-029import java.util.Collection;
-030import java.util.EnumSet;
-031import java.util.HashMap;
-032import java.util.Iterator;
-033import java.util.LinkedList;
-034import java.util.List;
-035import java.util.Map;
-036import java.util.Set;
-037import java.util.concurrent.Callable;
-038import 
java.util.concurrent.ExecutionException;
-039import java.util.concurrent.Future;
-040import java.util.concurrent.TimeUnit;
-041import 
java.util.concurrent.TimeoutException;
-042import 
java.util.concurrent.atomic.AtomicInteger;
-043import 
java.util.concurrent.atomic.AtomicReference;
-044import java.util.regex.Pattern;
-045import java.util.stream.Collectors;
-046import java.util.stream.Stream;
-047import 
org.apache.hadoop.conf.Configuration;
-048import 
org.apache.hadoop.hbase.Abortable;
-049import 
org.apache.hadoop.hbase.CacheEvictionStats;
-050import 
org.apache.hadoop.hbase.CacheEvictionStatsBuilder;
-051import 
org.apache.hadoop.hbase.ClusterMetrics.Option;
-052import 
org.apache.hadoop.hbase.ClusterStatus;
-053import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-054import 
org.apache.hadoop.hbase.HBaseConfiguration;
-055import 
org.apache.hadoop.hbase.HConstants;
-056import 
org.apache.hadoop.hbase.HRegionInfo;
-057import 
org.apache.hadoop.hbase.HRegionLocation;
-058import 
org.apache.hadoop.hbase.HTableDescriptor;
-059import 
org.apache.hadoop.hbase.MasterNotRunningException;
-060import 
org.apache.hadoop.hbase.MetaTableAccessor;
-061import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-062import 
org.apache.hadoop.hbase.NamespaceNotFoundException;
-063import 
org.apache.hadoop.hbase.NotServingRegionException;
-064import 
org.apache.hadoop.hbase.RegionLoad;
-065import 
org.apache.hadoop.hbase.RegionLocations;
-066import 
org.apache.hadoop.hbase.ServerName;
-067import 
org.apache.hadoop.hbase.TableExistsException;
-068import 
org.apache.hadoop.hbase.TableName;
-069import 
org.apache.hadoop.hbase.TableNotDisabledException;
-070import 
org.apache.hadoop.hbase.TableNotFoundException;
-071import 
org.apache.hadoop.hbase.UnknownRegionException;
-072import 
org.apache.hadoop.hbase.ZooKeeperConnectionException;
-073import 
org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil;
-074import 
org.apache.hadoop.hbase.client.replication.TableCFs;
-075import 
org.apache.hadoop.hbase.client.security.SecurityCapability;
-076import 
org.apache.hadoop.hbase.exceptions.TimeoutIOException;
-077import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
-078import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
-079import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-080import 
org.apache.hadoop.hbase.ipc.RpcControllerFactory;
-081import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-082import 
org.apache.hadoop.hbase.quotas.QuotaRetriever;
-083import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-084import 
org.apache.hadoop.hbase.regionserver.wal.FailedLogCloseException;
-085import 
org.apache.hadoop.hbase.replication.ReplicationException;
-086import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-087import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-088import 
org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
-089import 
org.apache.hadoop.hbase.snapshot.HBaseSnapshotException;
-090import 
org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
-091import 
org.apache.hadoop.hbase.snapshot.SnapshotCreationException;
-092import 
org.apache.hadoop.hbase.snapshot.UnknownSnapshotException;
-093import 
org.apache.hadoop.hbase.util.Addressing;
-094import 
org.apache.hadoop.hbase.util.Bytes;
-095import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-096import 
org.apache.hadoop.hbase.util.ForeignExceptionUtil;
-097import 
org.apache.hadoop.hbase.util.Pair;
-098import 
org.apache.hadoop.ipc.RemoteException;
-099import 
org.apache.hadoop.util.StringUtils;
-100import 
org.apache.yetus.audience.InterfaceAudience;
-101import 
org.apache.yetus.audience.InterfaceStability;
-102import org.slf4j.Logger;
-103import org.slf4j.LoggerFactory;
-104
-105import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-106import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-107import 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html 
b/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html
index 394abd6..ea801d2 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html
@@ -80,278 +80,290 @@
 072  }
 073  /**
 074   * Copy constructor
-075   * @param i
+075   * @param incrementToCopy increment to 
copy
 076   */
-077  public Increment(Increment i) {
-078this.row = i.getRow();
-079this.ts = i.getTimeStamp();
-080this.tr = i.getTimeRange();
-081
this.familyMap.putAll(i.getFamilyCellMap());
-082for (Map.EntryString, byte[] 
entry : i.getAttributesMap().entrySet()) {
-083  this.setAttribute(entry.getKey(), 
entry.getValue());
-084}
-085super.setPriority(i.getPriority());
-086  }
-087
-088  /**
-089   * Add the specified KeyValue to this 
operation.
-090   * @param cell individual Cell
-091   * @return this
-092   * @throws java.io.IOException e
-093   */
-094  public Increment add(Cell cell) throws 
IOException{
-095byte [] family = 
CellUtil.cloneFamily(cell);
-096ListCell list = 
getCellList(family);
-097//Checking that the row of the kv is 
the same as the put
-098if (!CellUtil.matchingRows(cell, 
this.row)) {
-099  throw new WrongRowIOException("The 
row in " + cell +
-100" doesn't match the original one 
" +  Bytes.toStringBinary(this.row));
-101}
-102list.add(cell);
-103return this;
-104  }
-105
-106  /**
-107   * Increment the column from the 
specific family with the specified qualifier
-108   * by the specified amount.
-109   * p
-110   * Overrides previous calls to 
addColumn for this family and qualifier.
-111   * @param family family name
-112   * @param qualifier column qualifier
-113   * @param amount amount to increment 
by
-114   * @return the Increment object
-115   */
-116  public Increment addColumn(byte [] 
family, byte [] qualifier, long amount) {
-117if (family == null) {
-118  throw new 
IllegalArgumentException("family cannot be null");
-119}
-120ListCell list = 
getCellList(family);
-121KeyValue kv = 
createPutKeyValue(family, qualifier, ts, Bytes.toBytes(amount));
-122list.add(kv);
-123return this;
-124  }
-125
-126  /**
-127   * Gets the TimeRange used for this 
increment.
-128   * @return TimeRange
-129   */
-130  public TimeRange getTimeRange() {
-131return this.tr;
-132  }
-133
-134  /**
-135   * Sets the TimeRange to be used on the 
Get for this increment.
-136   * p
-137   * This is useful for when you have 
counters that only last for specific
-138   * periods of time (ie. counters that 
are partitioned by time).  By setting
-139   * the range of valid times for this 
increment, you can potentially gain
-140   * some performance with a more optimal 
Get operation.
-141   * Be careful adding the time range to 
this class as you will update the old cell if the
-142   * time range doesn't include the 
latest cells.
-143   * p
-144   * This range is used as [minStamp, 
maxStamp).
-145   * @param minStamp minimum timestamp 
value, inclusive
-146   * @param maxStamp maximum timestamp 
value, exclusive
-147   * @throws IOException if invalid time 
range
-148   * @return this
-149   */
-150  public Increment setTimeRange(long 
minStamp, long maxStamp)
-151  throws IOException {
-152tr = new TimeRange(minStamp, 
maxStamp);
-153return this;
-154  }
-155
-156  @Override
-157  public Increment setTimestamp(long 
timestamp) {
-158super.setTimestamp(timestamp);
+077  public Increment(Increment 
incrementToCopy) {
+078super(incrementToCopy);
+079this.tr = 
incrementToCopy.getTimeRange();
+080  }
+081
+082  /**
+083   * Construct the Increment with user 
defined data. NOTED:
+084   * 1) all cells in the familyMap must 
have the DataType.Put
+085   * 2) the row of each cell must be same 
with passed row.
+086   * @param row row. CAN'T be null
+087   * @param ts timestamp
+088   * @param familyMap the map to collect 
all cells internally. CAN'T be null
+089   */
+090  public Increment(byte[] row, long ts, 
NavigableMapbyte [], ListCell familyMap) {
+091super(row, ts, familyMap);
+092  }
+093
+094  /**
+095   * Add the specified KeyValue to this 
operation.
+096   * @param cell individual Cell
+097   * @return this
+098   * @throws java.io.IOException e
+099   */
+100  public Increment add(Cell cell) throws 
IOException{
+101byte [] family = 
CellUtil.cloneFamily(cell);
+102ListCell list = 
getCellList(family);
+103//Checking that the row of the kv is 
the same as the put
+104if (!CellUtil.matchingRows(cell, 
this.row)) {
+105  throw new WrongRowIOException("The 
row in " + cell +
+106" doesn't match the 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html 
b/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html
index 51d92c2..86fc15e 100644
--- 
a/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html
+++ 
b/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html
@@ -44,2578 +44,2580 @@
 036import java.util.Iterator;
 037import java.util.List;
 038
-039import com.google.protobuf.ByteString;
-040import org.apache.commons.logging.Log;
-041import 
org.apache.commons.logging.LogFactory;
-042import org.apache.hadoop.hbase.Cell;
-043import 
org.apache.hadoop.hbase.CellComparator;
-044import 
org.apache.hadoop.hbase.KeyValue;
-045import 
org.apache.hadoop.io.RawComparator;
-046import 
org.apache.hadoop.io.WritableComparator;
-047import 
org.apache.hadoop.io.WritableUtils;
-048import 
org.apache.yetus.audience.InterfaceAudience;
-049import sun.misc.Unsafe;
-050
-051import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-052import 
org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;
+039import org.apache.hadoop.hbase.Cell;
+040import 
org.apache.hadoop.hbase.CellComparator;
+041import 
org.apache.hadoop.hbase.KeyValue;
+042import 
org.apache.hadoop.io.RawComparator;
+043import 
org.apache.hadoop.io.WritableComparator;
+044import 
org.apache.hadoop.io.WritableUtils;
+045import 
org.apache.yetus.audience.InterfaceAudience;
+046import org.slf4j.Logger;
+047import org.slf4j.LoggerFactory;
+048
+049import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
+050import 
org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;
+051
+052import com.google.protobuf.ByteString;
 053
-054/**
-055 * Utility class that handles byte 
arrays, conversions to/from other types,
-056 * comparisons, hash code generation, 
manufacturing keys for HashMaps or
-057 * HashSets, and can be used as key in 
maps or trees.
-058 */
-059@SuppressWarnings("restriction")
-060@InterfaceAudience.Public
-061@edu.umd.cs.findbugs.annotations.SuppressWarnings(
-062
value="EQ_CHECK_FOR_OPERAND_NOT_COMPATIBLE_WITH_THIS",
-063justification="It has been like this 
forever")
-064public class Bytes implements 
ComparableBytes {
-065
-066  // Using the charset canonical name for 
String/byte[] conversions is much
-067  // more efficient due to use of cached 
encoders/decoders.
-068  private static final String UTF8_CSN = 
StandardCharsets.UTF_8.name();
-069
-070  //HConstants.EMPTY_BYTE_ARRAY should be 
updated if this changed
-071  private static final byte [] 
EMPTY_BYTE_ARRAY = new byte [0];
-072
-073  private static final Log LOG = 
LogFactory.getLog(Bytes.class);
+054import sun.misc.Unsafe;
+055
+056/**
+057 * Utility class that handles byte 
arrays, conversions to/from other types,
+058 * comparisons, hash code generation, 
manufacturing keys for HashMaps or
+059 * HashSets, and can be used as key in 
maps or trees.
+060 */
+061@SuppressWarnings("restriction")
+062@InterfaceAudience.Public
+063@edu.umd.cs.findbugs.annotations.SuppressWarnings(
+064
value="EQ_CHECK_FOR_OPERAND_NOT_COMPATIBLE_WITH_THIS",
+065justification="It has been like this 
forever")
+066public class Bytes implements 
ComparableBytes {
+067
+068  // Using the charset canonical name for 
String/byte[] conversions is much
+069  // more efficient due to use of cached 
encoders/decoders.
+070  private static final String UTF8_CSN = 
StandardCharsets.UTF_8.name();
+071
+072  //HConstants.EMPTY_BYTE_ARRAY should be 
updated if this changed
+073  private static final byte [] 
EMPTY_BYTE_ARRAY = new byte [0];
 074
-075  /**
-076   * Size of boolean in bytes
-077   */
-078  public static final int SIZEOF_BOOLEAN 
= Byte.SIZE / Byte.SIZE;
-079
-080  /**
-081   * Size of byte in bytes
-082   */
-083  public static final int SIZEOF_BYTE = 
SIZEOF_BOOLEAN;
-084
-085  /**
-086   * Size of char in bytes
-087   */
-088  public static final int SIZEOF_CHAR = 
Character.SIZE / Byte.SIZE;
-089
-090  /**
-091   * Size of double in bytes
-092   */
-093  public static final int SIZEOF_DOUBLE = 
Double.SIZE / Byte.SIZE;
-094
-095  /**
-096   * Size of float in bytes
-097   */
-098  public static final int SIZEOF_FLOAT = 
Float.SIZE / Byte.SIZE;
-099
-100  /**
-101   * Size of int in bytes
-102   */
-103  public static final int SIZEOF_INT = 
Integer.SIZE / Byte.SIZE;
-104
-105  /**
-106   * Size of long in bytes
-107   */
-108  public static final int SIZEOF_LONG = 
Long.SIZE / Byte.SIZE;
-109
-110  /**
-111   * Size of short in bytes
-112   */
-113  public static final int SIZEOF_SHORT = 
Short.SIZE / Byte.SIZE;
-114
-115  /**
-116   * Mask to apply to a long to reveal 
the lower int only. Use like 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html
index 447ee2d..f943678 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html
@@ -99,314 +99,316 @@
 091  /**
 092   * @return The filter serialized using 
pb
 093   */
-094  public byte [] toByteArray() {
-095FilterProtos.KeyOnlyFilter.Builder 
builder =
-096  
FilterProtos.KeyOnlyFilter.newBuilder();
-097builder.setLenAsVal(this.lenAsVal);
-098return 
builder.build().toByteArray();
-099  }
-100
-101  /**
-102   * @param pbBytes A pb serialized 
{@link KeyOnlyFilter} instance
-103   * @return An instance of {@link 
KeyOnlyFilter} made from codebytes/code
-104   * @throws DeserializationException
-105   * @see #toByteArray
-106   */
-107  public static KeyOnlyFilter 
parseFrom(final byte [] pbBytes)
-108  throws DeserializationException {
-109FilterProtos.KeyOnlyFilter proto;
-110try {
-111  proto = 
FilterProtos.KeyOnlyFilter.parseFrom(pbBytes);
-112} catch 
(InvalidProtocolBufferException e) {
-113  throw new 
DeserializationException(e);
-114}
-115return new 
KeyOnlyFilter(proto.getLenAsVal());
-116  }
-117
-118  /**
-119   * @param o the other filter to compare 
with
-120   * @return true if and only if the 
fields of the filter that are serialized
-121   * are equal to the corresponding 
fields in other.  Used for testing.
-122   */
-123  boolean areSerializedFieldsEqual(Filter 
o) {
-124if (o == this) return true;
-125if (!(o instanceof KeyOnlyFilter)) 
return false;
-126
-127KeyOnlyFilter other = 
(KeyOnlyFilter)o;
-128return this.lenAsVal == 
other.lenAsVal;
-129  }
-130
-131  static class KeyOnlyCell implements 
Cell {
-132private Cell cell;
-133private boolean lenAsVal;
-134
-135public KeyOnlyCell(Cell c, boolean 
lenAsVal) {
-136  this.cell = c;
-137  this.lenAsVal = lenAsVal;
-138}
-139
-140@Override
-141public byte[] getRowArray() {
-142  return cell.getRowArray();
-143}
-144
-145@Override
-146public int getRowOffset() {
-147  return cell.getRowOffset();
-148}
-149
-150@Override
-151public short getRowLength() {
-152  return cell.getRowLength();
-153}
-154
-155@Override
-156public byte[] getFamilyArray() {
-157  return cell.getFamilyArray();
-158}
-159
-160@Override
-161public int getFamilyOffset() {
-162  return cell.getFamilyOffset();
-163}
-164
-165@Override
-166public byte getFamilyLength() {
-167  return cell.getFamilyLength();
-168}
-169
-170@Override
-171public byte[] getQualifierArray() {
-172  return cell.getQualifierArray();
-173}
-174
-175@Override
-176public int getQualifierOffset() {
-177  return cell.getQualifierOffset();
-178}
-179
-180@Override
-181public int getQualifierLength() {
-182  return cell.getQualifierLength();
-183}
-184
-185@Override
-186public long getTimestamp() {
-187  return cell.getTimestamp();
-188}
-189
-190@Override
-191public byte getTypeByte() {
-192  return cell.getTypeByte();
-193}
-194
-195@Override
-196public long getSequenceId() {
-197  return 0;
-198}
-199
-200@Override
-201public byte[] getValueArray() {
-202  if (lenAsVal) {
-203return 
Bytes.toBytes(cell.getValueLength());
-204  } else {
-205return 
HConstants.EMPTY_BYTE_ARRAY;
-206  }
-207}
-208
-209@Override
-210public int getValueOffset() {
-211  return 0;
-212}
-213
-214@Override
-215public int getValueLength() {
-216  if (lenAsVal) {
-217return Bytes.SIZEOF_INT;
-218  } else {
-219return 0;
-220  }
-221}
-222
-223@Override
-224public byte[] getTagsArray() {
-225  return 
HConstants.EMPTY_BYTE_ARRAY;
-226}
-227
-228@Override
-229public int getTagsOffset() {
-230  return 0;
-231}
-232
-233@Override
-234public int getTagsLength() {
-235  return 0;
-236}
-237  }
-238
-239  static class KeyOnlyByteBufferCell 
extends ByteBufferCell {
-240private ByteBufferCell cell;
-241private boolean lenAsVal;
-242
-243public 
KeyOnlyByteBufferCell(ByteBufferCell c, boolean lenAsVal) {
-244  this.cell = c;
-245  this.lenAsVal = lenAsVal;
-246}
-247
-248@Override
-249public byte[] getRowArray() {
-250  return cell.getRowArray();
-251}
-252
-253@Override
-254public int getRowOffset() {
-255  return cell.getRowOffset();
-256}
-257
-258@Override
-259public short getRowLength() {
-260  return 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
index f1a2443..a469e93 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html
@@ -1350,415 +1350,415 @@
 1342return delete;
 1343  }
 1344
-1345  public static Put 
makeBarrierPut(byte[] encodedRegionName, long seq, byte[] tableName) {
-1346byte[] seqBytes = 
Bytes.toBytes(seq);
-1347return new Put(encodedRegionName)
-1348
.addImmutable(HConstants.REPLICATION_BARRIER_FAMILY, seqBytes, seqBytes)
-1349
.addImmutable(HConstants.REPLICATION_META_FAMILY, tableNameCq, tableName);
-1350  }
-1351
-1352
-1353  public static Put 
makeDaughterPut(byte[] encodedRegionName, byte[] value) {
-1354return new 
Put(encodedRegionName).addImmutable(HConstants.REPLICATION_META_FAMILY,
-1355daughterNameCq, value);
-1356  }
-1357
-1358  public static Put makeParentPut(byte[] 
encodedRegionName, byte[] value) {
-1359return new 
Put(encodedRegionName).addImmutable(HConstants.REPLICATION_META_FAMILY,
-1360parentNameCq, value);
-1361  }
-1362
-1363  /**
-1364   * Adds split daughters to the Put
-1365   */
-1366  public static Put 
addDaughtersToPut(Put put, RegionInfo splitA, RegionInfo splitB) {
-1367if (splitA != null) {
-1368  put.addImmutable(
-1369HConstants.CATALOG_FAMILY, 
HConstants.SPLITA_QUALIFIER, RegionInfo.toByteArray(splitA));
-1370}
-1371if (splitB != null) {
-1372  put.addImmutable(
-1373HConstants.CATALOG_FAMILY, 
HConstants.SPLITB_QUALIFIER, RegionInfo.toByteArray(splitB));
-1374}
-1375return put;
-1376  }
-1377
-1378  /**
-1379   * Put the passed 
codeputs/code to the codehbase:meta/code 
table.
-1380   * Non-atomic for multi puts.
-1381   * @param connection connection we're 
using
-1382   * @param puts Put to add to 
hbase:meta
-1383   * @throws IOException
-1384   */
-1385  public static void 
putToMetaTable(final Connection connection, final Put... puts)
-1386throws IOException {
-1387put(getMetaHTable(connection), 
Arrays.asList(puts));
-1388  }
-1389
-1390  /**
-1391   * @param t Table to use (will be 
closed when done).
-1392   * @param puts puts to make
-1393   * @throws IOException
-1394   */
-1395  private static void put(final Table t, 
final ListPut puts) throws IOException {
-1396try {
-1397  if (METALOG.isDebugEnabled()) {
-1398
METALOG.debug(mutationsToString(puts));
-1399  }
-1400  t.put(puts);
-1401} finally {
-1402  t.close();
-1403}
-1404  }
-1405
-1406  /**
-1407   * Put the passed 
codeps/code to the codehbase:meta/code table.
-1408   * @param connection connection we're 
using
-1409   * @param ps Put to add to 
hbase:meta
-1410   * @throws IOException
-1411   */
-1412  public static void 
putsToMetaTable(final Connection connection, final ListPut ps)
-1413throws IOException {
-1414Table t = 
getMetaHTable(connection);
-1415try {
-1416  if (METALOG.isDebugEnabled()) {
-1417
METALOG.debug(mutationsToString(ps));
-1418  }
-1419  t.put(ps);
-1420} finally {
-1421  t.close();
-1422}
-1423  }
-1424
-1425  /**
-1426   * Delete the passed 
coded/code from the codehbase:meta/code 
table.
-1427   * @param connection connection we're 
using
-1428   * @param d Delete to add to 
hbase:meta
-1429   * @throws IOException
-1430   */
-1431  static void deleteFromMetaTable(final 
Connection connection, final Delete d)
-1432throws IOException {
-1433ListDelete dels = new 
ArrayList(1);
-1434dels.add(d);
-1435deleteFromMetaTable(connection, 
dels);
-1436  }
-1437
-1438  /**
-1439   * Delete the passed 
codedeletes/code from the codehbase:meta/code 
table.
-1440   * @param connection connection we're 
using
-1441   * @param deletes Deletes to add to 
hbase:meta  This list should support #remove.
-1442   * @throws IOException
-1443   */
-1444  public static void 
deleteFromMetaTable(final Connection connection, final ListDelete 
deletes)
-1445throws IOException {
-1446Table t = 
getMetaHTable(connection);
-1447try {
-1448  if (METALOG.isDebugEnabled()) {
-1449
METALOG.debug(mutationsToString(deletes));
-1450  }
-1451  t.delete(deletes);
-1452} finally {
-1453  t.close();
-1454}
-1455  }
-1456
-1457  /**
-1458   * Deletes some replica columns 
corresponding to replicas for the passed rows
-1459   * @param metaRows rows in 
hbase:meta
-1460   * @param replicaIndexToDeleteFrom the 
replica ID we would start deleting from
-1461   * @param 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html
index 7c59e27..c904c56 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html
@@ -119,4048 +119,4054 @@
 111import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
 112import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
 113import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearRegionBlockCacheRequest;
-114import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
-115import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
-116import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
-117import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
-118import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest;
-119import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse;
-120import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest;
-121import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest;
-122import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-123import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest;
-124import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse;
-125import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
-126import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
-127import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
-128import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-129import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos;
-130import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
-131import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
-132import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-133import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-134import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-135import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest;
-136import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-137import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-138import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest;
-139import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse;
-140import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
-141import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
-142import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
-143import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse;
-144import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest;
-145import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest;
-146import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse;
-147import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
-148import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
-149import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
-150import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
-151import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest;
-152import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse;
-153import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest;
-154import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest;
-155import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest;

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-13 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/devapidocs/org/apache/hadoop/hbase/client/HTable.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/HTable.html 
b/devapidocs/org/apache/hadoop/hbase/client/HTable.html
index 380c1d7..a4d8801 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/HTable.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/HTable.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":42,"i8":10,"i9":42,"i10":10,"i11":10,"i12":42,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":9,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":9,"i33":10,"i34":10,"i35":10,"i36":9,"i37":10,"i38":42,"i39":10,"i40":10,"i41":42,"i42":10,"i43":10,"i44":42,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":42,"i51":42,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":42,"i61":42,"i62":42,"i63":42,"i64":10,"i65":10,"i66":9};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":42,"i7":42,"i8":42,"i9":10,"i10":42,"i11":42,"i12":42,"i13":42,"i14":42,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":9,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":9,"i34":10,"i35":10,"i36":10,"i37":9,"i38":10,"i39":42,"i40":10,"i41":10,"i42":42,"i43":10,"i44":10,"i45":42,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":42,"i52":42,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":42,"i62":42,"i63":42,"i64":42,"i65":10,"i66":10,"i67":9};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
 
 
 PrevClass
-NextClass
+NextClass
 
 
 Frames
@@ -74,7 +74,7 @@ var activeTableTab = "activeTableTab";
 
 
 Summary:
-Nested|
+Nested|
 Field|
 Constr|
 Method
@@ -115,7 +115,7 @@ var activeTableTab = "activeTableTab";
 
 @InterfaceAudience.Private
  @InterfaceStability.Stable
-public class HTable
+public class HTable
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements Table
 An implementation of Table. 
Used to communicate with a single HBase table.
@@ -147,6 +147,32 @@ implements 
 
 
+
+
+
+
+
+Nested Class Summary
+
+Nested Classes
+
+Modifier and Type
+Class and Description
+
+
+private class
+HTable.CheckAndMutateBuilderImpl
+
+
+
+
+
+
+Nested classes/interfaces inherited from 
interfaceorg.apache.hadoop.hbase.client.Table
+Table.CheckAndMutateBuilder
+
+
+
 
 
 
@@ -336,8 +362,7 @@ implements Deletedelete)
-Atomically checks if a row/family/qualifier value matches 
the expected
- value.
+Deprecated.
 
 
 
@@ -359,11 +384,17 @@ implements CompareOperatorop,
   byte[]value,
   Deletedelete)
-Atomically checks if a row/family/qualifier value matches 
the expected
- value.
+Deprecated.
 
 
 
+Table.CheckAndMutateBuilder
+checkAndMutate(byte[]row,
+  byte[]family)
+Atomically checks if a row/family/qualifier value matches 
the expected value.
+
+
+
 boolean
 checkAndMutate(byte[]row,
   byte[]family,
@@ -374,7 +405,7 @@ implements Deprecated.
 
 
-
+
 boolean
 checkAndMutate(byte[]row,
   byte[]family,
@@ -382,21 +413,20 @@ implements CompareOperatorop,
   byte[]value,
   RowMutationsrm)
-Atomically checks if a row/family/qualifier value matches 
the expected value.
+Deprecated.
 
 
-
+
 boolean
 checkAndPut(byte[]row,
byte[]family,
byte[]qualifier,
byte[]value,
Putput)
-Atomically checks if a row/family/qualifier value matches 
the expected
- value.
+Deprecated.
 
 
-
+
 boolean
 checkAndPut(byte[]row,
byte[]family,
@@ -407,7 +437,7 @@ implements Deprecated.
 
 
-
+
 boolean
 checkAndPut(byte[]row,
byte[]family,
@@ -415,30 +445,29 @@ implements CompareOperatorop,
byte[]value,
Putput)
-Atomically checks if a row/family/qualifier value matches 
the expected
- value.
+Deprecated.
 
 
-
+
 void
 clearRegionCache()
 Explicitly clears the region cache to fetch the latest 
value from META.
 
 
-
+
 void
 close()
 Releases any resources held or pending changes in internal 
buffers.
 
 
-
+
 CoprocessorRpcChannel
 coprocessorService(byte[]row)
 Creates and returns a RpcChannel instance 
connected to the
  table region containing the specified row.
 
 
-
+
 T extends 
com.google.protobuf.Service,Rhttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Mapbyte[],R
 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html 
b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
index a5af31c..a668610 100644
--- a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
+++ b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10};
+var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
 
 
 All Implemented Interfaces:
-Cell, SettableSequenceId
+http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true;
 title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell
 
 
 Direct Known Subclasses:
@@ -131,7 +131,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class PrivateCellUtil.FirstOnRowColCell
+private static class PrivateCellUtil.FirstOnRowColCell
 extends PrivateCellUtil.FirstOnRowCell
 
 
@@ -156,26 +156,44 @@ extends fArray
 
 
+private static long
+FIXED_HEAPSIZE
+
+
 private byte
 flength
 
-
+
 private int
 foffset
 
-
+
 private byte[]
 qArray
 
-
+
 private int
 qlength
 
-
+
 private int
 qoffset
 
 
+
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell
+CELL_NOT_BASED_ON_CHUNK
+
+
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.RawCell
+MAX_TAGS_LENGTH
+
 
 
 
@@ -244,6 +262,10 @@ extends int
 getQualifierOffset()
 
+
+long
+heapSize()
+
 
 
 
@@ -257,7 +279,7 @@ extends PrivateCellUtil.EmptyCell
-getSequenceId,
 getTagsArray,
 getTagsLength,
 getTagsOffset,
 getValueArray,
 getValueLength,
 getValueOffset,
 setSequenceId
+getSequenceId,
 getTagsArray,
 getTagsLength,
 getTagsOffset,
 getValueArray,
 getValueLength,
 getValueOffset,
 setSequenceId,
 setTimestamp, setTimestamp
 
 
 
@@ -266,6 +288,20 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
 title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
 title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
 title="class or interface in java.lang">notify, http://docs.oracle.com/javase/8/docs/api/java/lang
 /Object.html?is-external=true#notifyAll--" title="class or interface in 
java.lang">notifyAll, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
 title="class or interface in java.lang">toString, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--;
 title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-;
 title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-;
 title="class or interface in java.lang">wait
 
+
+
+
+
+Methods inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell
+deepClone,
 getChunkId,
 getSerializedSize,
 write,
 write
+
+
+
+
+
+Methods inherited from interfaceorg.apache.hadoop.hbase.RawCell
+checkForTagsLength,
 cloneTags,
 getTag,
 getTags
+
 
 
 
@@ -280,13 +316,22 @@ extends 
+
+
+
+
+FIXED_HEAPSIZE
+private static finallong FIXED_HEAPSIZE
+
+
 
 
 
 
 
 fArray
-private finalbyte[] fArray
+private finalbyte[] fArray
 
 
 
@@ -295,7 +340,7 @@ extends 
 
 foffset
-private finalint foffset
+private finalint foffset
 
 
 
@@ -304,7 +349,7 @@ extends 
 
 flength
-private finalbyte flength
+private finalbyte flength
 
 
 
@@ -313,7 +358,7 @@ extends 
 
 qArray
-private finalbyte[] qArray
+private finalbyte[] qArray
 
 
 
@@ -322,7 +367,7 @@ extends 
 
 qoffset
-private finalint qoffset
+private finalint qoffset
 
 
 
@@ -331,7 +376,7 @@ extends 
 
 qlength
-private finalint qlength
+private finalint qlength
 
 
 
@@ -348,7 +393,7 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-06 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html
index 4c37cbe..cfb8ee4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html
@@ -228,837 +228,853 @@
 220   * of candidates. If you remove all the 
candidates then the compaction will be canceled.
 221   * pSupports Coprocessor 
'bypass' -- 'bypass' is how this method indicates that it changed
 222   * the passed in 
codecandidates/code.
-223   * @param c the environment provided by 
the region server
-224   * @param store the store where 
compaction is being requested
-225   * @param candidates the store files 
currently available for compaction
-226   * @param tracker tracker used to track 
the life cycle of a compaction
-227   */
-228  default void 
preCompactSelection(ObserverContextRegionCoprocessorEnvironment c, 
Store store,
-229  List? extends StoreFile 
candidates, CompactionLifeCycleTracker tracker)
-230  throws IOException {}
-231
-232  /**
-233   * Called after the {@link StoreFile}s 
to compact have been selected from the available
-234   * candidates.
-235   * @param c the environment provided by 
the region server
-236   * @param store the store being 
compacted
-237   * @param selected the store files 
selected to compact
-238   * @param tracker tracker used to track 
the life cycle of a compaction
-239   * @param request the requested 
compaction
-240   */
-241  default void 
postCompactSelection(ObserverContextRegionCoprocessorEnvironment c, 
Store store,
-242  List? extends StoreFile 
selected, CompactionLifeCycleTracker tracker,
-243  CompactionRequest request) {}
-244
-245  /**
-246   * Called before we open store scanner 
for compaction. You can use the {@code options} to change max
-247   * versions and TTL for the scanner 
being opened.
-248   * @param c the environment provided by 
the region server
-249   * @param store the store being 
compacted
-250   * @param scanType type of Scan
-251   * @param options used to change max 
versions and TTL for the scanner being opened
-252   * @param tracker tracker used to track 
the life cycle of a compaction
-253   * @param request the requested 
compaction
-254   */
-255  default void 
preCompactScannerOpen(ObserverContextRegionCoprocessorEnvironment c, 
Store store,
-256  ScanType scanType, ScanOptions 
options, CompactionLifeCycleTracker tracker,
-257  CompactionRequest request) throws 
IOException {}
-258
-259  /**
-260   * Called prior to writing the {@link 
StoreFile}s selected for compaction into a new
-261   * {@code StoreFile}.
-262   * p
-263   * To override or modify the compaction 
process, implementing classes can wrap the provided
-264   * {@link InternalScanner} with a 
custom implementation that is returned from this method. The
-265   * custom scanner can then inspect 
{@link org.apache.hadoop.hbase.Cell}s from the wrapped scanner,
-266   * applying its own policy to what gets 
written.
-267   * @param c the environment provided by 
the region server
-268   * @param store the store being 
compacted
-269   * @param scanner the scanner over 
existing data used in the store file rewriting
-270   * @param scanType type of Scan
-271   * @param tracker tracker used to track 
the life cycle of a compaction
-272   * @param request the requested 
compaction
-273   * @return the scanner to use during 
compaction. Should not be {@code null} unless the
-274   * implementation is writing 
new store files on its own.
-275   */
-276  default InternalScanner 
preCompact(ObserverContextRegionCoprocessorEnvironment c, Store 
store,
-277  InternalScanner scanner, ScanType 
scanType, CompactionLifeCycleTracker tracker,
-278  CompactionRequest request) throws 
IOException {
-279return scanner;
-280  }
-281
-282  /**
-283   * Called after compaction has 
completed and the new store file has been moved in to place.
-284   * @param c the environment provided by 
the region server
-285   * @param store the store being 
compacted
-286   * @param resultFile the new store file 
written out during compaction
-287   * @param tracker used to track the 
life cycle of a compaction
-288   * @param request the requested 
compaction
-289   */
-290  default void 
postCompact(ObserverContextRegionCoprocessorEnvironment c, Store 
store,
-291  StoreFile resultFile, 
CompactionLifeCycleTracker tracker, CompactionRequest request)
-292  throws IOException {}
-293
-294  /**
-295   * Called before the region is reported 
as closed to the master.
-296   * @param c the environment provided by 
the region server
-297   * @param 

[23/51] [partial] hbase-site git commit: Published site at .

2017-12-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
index 3edfbef..9707b2c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
@@ -2459,5936 +2459,5935 @@
 2451  }
 2452
 2453  for (HStore s : storesToFlush) {
-2454MemStoreSize flushableSize = 
s.getFlushableSize();
-2455
totalSizeOfFlushableStores.incMemStoreSize(flushableSize);
-2456
storeFlushCtxs.put(s.getColumnFamilyDescriptor().getName(),
-2457  
s.createFlushContext(flushOpSeqId, tracker));
-2458// for writing stores to WAL
-2459
committedFiles.put(s.getColumnFamilyDescriptor().getName(), null);
-2460
storeFlushableSize.put(s.getColumnFamilyDescriptor().getName(), 
flushableSize);
-2461  }
-2462
-2463  // write the snapshot start to 
WAL
-2464  if (wal != null  
!writestate.readOnly) {
-2465FlushDescriptor desc = 
ProtobufUtil.toFlushDescriptor(FlushAction.START_FLUSH,
-2466getRegionInfo(), 
flushOpSeqId, committedFiles);
-2467// No sync. Sync is below where 
no updates lock and we do FlushAction.COMMIT_FLUSH
-2468WALUtil.writeFlushMarker(wal, 
this.getReplicationScope(), getRegionInfo(), desc, false,
-2469mvcc);
-2470  }
-2471
-2472  // Prepare flush (take a 
snapshot)
-2473  for (StoreFlushContext flush : 
storeFlushCtxs.values()) {
-2474flush.prepare();
-2475  }
-2476} catch (IOException ex) {
-2477  doAbortFlushToWAL(wal, 
flushOpSeqId, committedFiles);
-2478  throw ex;
-2479} finally {
-2480  
this.updatesLock.writeLock().unlock();
-2481}
-2482String s = "Finished memstore 
snapshotting " + this + ", syncing WAL and waiting on mvcc, " +
-2483"flushsize=" + 
totalSizeOfFlushableStores;
-2484status.setStatus(s);
-2485doSyncOfUnflushedWALChanges(wal, 
getRegionInfo());
-2486return new 
PrepareFlushResult(storeFlushCtxs, committedFiles, storeFlushableSize, 
startTime,
-2487flushOpSeqId, flushedSeqId, 
totalSizeOfFlushableStores);
-2488  }
-2489
-2490  /**
-2491   * Utility method broken out of 
internalPrepareFlushCache so that method is smaller.
-2492   */
-2493  private void 
logFatLineOnFlush(CollectionHStore storesToFlush, long sequenceId) {
-2494if (!LOG.isInfoEnabled()) {
-2495  return;
-2496}
-2497// Log a fat line detailing what is 
being flushed.
-2498StringBuilder perCfExtras = null;
-2499if (!isAllFamilies(storesToFlush)) 
{
-2500  perCfExtras = new 
StringBuilder();
-2501  for (HStore store: storesToFlush) 
{
-2502perCfExtras.append("; 
").append(store.getColumnFamilyName());
-2503perCfExtras.append("=")
-2504
.append(StringUtils.byteDesc(store.getFlushableSize().getDataSize()));
-2505  }
-2506}
-2507LOG.info("Flushing " + + 
storesToFlush.size() + "/" + stores.size() +
-2508" column families, memstore=" + 
StringUtils.byteDesc(this.memstoreDataSize.get()) +
-2509((perCfExtras != null  
perCfExtras.length()  0)? perCfExtras.toString(): "") +
-2510((wal != null) ? "" : "; WAL is 
null, using passed sequenceid=" + sequenceId));
-2511  }
-2512
-2513  private void doAbortFlushToWAL(final 
WAL wal, final long flushOpSeqId,
-2514  final Mapbyte[], 
ListPath committedFiles) {
-2515if (wal == null) return;
-2516try {
-2517  FlushDescriptor desc = 
ProtobufUtil.toFlushDescriptor(FlushAction.ABORT_FLUSH,
-2518  getRegionInfo(), flushOpSeqId, 
committedFiles);
-2519  WALUtil.writeFlushMarker(wal, 
this.getReplicationScope(), getRegionInfo(), desc, false,
-2520  mvcc);
-2521} catch (Throwable t) {
-2522  LOG.warn("Received unexpected 
exception trying to write ABORT_FLUSH marker to WAL:" +
-2523  
StringUtils.stringifyException(t));
-2524  // ignore this since we will be 
aborting the RS with DSE.
-2525}
-2526// we have called 
wal.startCacheFlush(), now we have to abort it
-2527
wal.abortCacheFlush(this.getRegionInfo().getEncodedNameAsBytes());
-2528  }
-2529
-2530  /**
-2531   * Sync unflushed WAL changes. See 
HBASE-8208 for details
-2532   */
-2533  private static void 
doSyncOfUnflushedWALChanges(final WAL wal, final RegionInfo hri)
-2534  throws IOException {
-2535if (wal == null) {
-2536  return;
-2537}
-2538try {
-2539  wal.sync(); // ensure that flush 
marker is sync'ed
-2540} catch (IOException ioe) {
-2541  

[23/51] [partial] hbase-site git commit: Published site at .

2017-11-30 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html
index 11bfb15..915e78a 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html
@@ -782,557 +782,562 @@
 774  /**
 775   * Default value of {@link 
#HBASE_CLIENT_RETRIES_NUMBER}.
 776   */
-777  public static final int 
DEFAULT_HBASE_CLIENT_RETRIES_NUMBER = 35;
+777  public static final int 
DEFAULT_HBASE_CLIENT_RETRIES_NUMBER = 10;
 778
-779  /**
-780   * Parameter name to set the default 
scanner caching for all clients.
-781   */
-782  public static final String 
HBASE_CLIENT_SCANNER_CACHING = "hbase.client.scanner.caching";
+779  public static final String 
HBASE_CLIENT_SERVERSIDE_RETRIES_MULTIPLIER =
+780  
"hbase.client.serverside.retries.multiplier";
+781
+782  public static final int 
DEFAULT_HBASE_CLIENT_SERVERSIDE_RETRIES_MULTIPLIER = 3;
 783
 784  /**
-785   * Default value for {@link 
#HBASE_CLIENT_SCANNER_CACHING}
+785   * Parameter name to set the default 
scanner caching for all clients.
 786   */
-787  public static final int 
DEFAULT_HBASE_CLIENT_SCANNER_CACHING = Integer.MAX_VALUE;
+787  public static final String 
HBASE_CLIENT_SCANNER_CACHING = "hbase.client.scanner.caching";
 788
 789  /**
-790   * Parameter name for number of rows 
that will be fetched when calling next on
-791   * a scanner if it is not served from 
memory. Higher caching values will
-792   * enable faster scanners but will eat 
up more memory and some calls of next
-793   * may take longer and longer times 
when the cache is empty.
-794   */
-795  public static final String 
HBASE_META_SCANNER_CACHING = "hbase.meta.scanner.caching";
-796
-797  /**
-798   * Default value of {@link 
#HBASE_META_SCANNER_CACHING}.
+790   * Default value for {@link 
#HBASE_CLIENT_SCANNER_CACHING}
+791   */
+792  public static final int 
DEFAULT_HBASE_CLIENT_SCANNER_CACHING = Integer.MAX_VALUE;
+793
+794  /**
+795   * Parameter name for number of rows 
that will be fetched when calling next on
+796   * a scanner if it is not served from 
memory. Higher caching values will
+797   * enable faster scanners but will eat 
up more memory and some calls of next
+798   * may take longer and longer times 
when the cache is empty.
 799   */
-800  public static final int 
DEFAULT_HBASE_META_SCANNER_CACHING = 100;
+800  public static final String 
HBASE_META_SCANNER_CACHING = "hbase.meta.scanner.caching";
 801
 802  /**
-803   * Parameter name for number of 
versions, kept by meta table.
+803   * Default value of {@link 
#HBASE_META_SCANNER_CACHING}.
 804   */
-805  public static final String 
HBASE_META_VERSIONS = "hbase.meta.versions";
+805  public static final int 
DEFAULT_HBASE_META_SCANNER_CACHING = 100;
 806
 807  /**
-808   * Default value of {@link 
#HBASE_META_VERSIONS}.
+808   * Parameter name for number of 
versions, kept by meta table.
 809   */
-810  public static final int 
DEFAULT_HBASE_META_VERSIONS = 3;
+810  public static final String 
HBASE_META_VERSIONS = "hbase.meta.versions";
 811
 812  /**
-813   * Parameter name for number of 
versions, kept by meta table.
+813   * Default value of {@link 
#HBASE_META_VERSIONS}.
 814   */
-815  public static final String 
HBASE_META_BLOCK_SIZE = "hbase.meta.blocksize";
+815  public static final int 
DEFAULT_HBASE_META_VERSIONS = 3;
 816
 817  /**
-818   * Default value of {@link 
#HBASE_META_BLOCK_SIZE}.
+818   * Parameter name for number of 
versions, kept by meta table.
 819   */
-820  public static final int 
DEFAULT_HBASE_META_BLOCK_SIZE = 8 * 1024;
+820  public static final String 
HBASE_META_BLOCK_SIZE = "hbase.meta.blocksize";
 821
 822  /**
-823   * Parameter name for unique identifier 
for this {@link org.apache.hadoop.conf.Configuration}
-824   * instance. If there are two or more 
{@link org.apache.hadoop.conf.Configuration} instances that,
-825   * for all intents and purposes, are 
the same except for their instance ids, then they will not be
-826   * able to share the same 
org.apache.hadoop.hbase.client.HConnection instance. On the other hand,
-827   * even if the instance ids are the 
same, it could result in non-shared
-828   * 
org.apache.hadoop.hbase.client.HConnection instances if some of the other 
connection parameters
-829   * differ.
-830   */
-831  public static final String 
HBASE_CLIENT_INSTANCE_ID = "hbase.client.instance.id";
-832
-833  /**
-834   * The client scanner timeout period in 
milliseconds.
+823   * Default value of {@link 
#HBASE_META_BLOCK_SIZE}.
+824   */
+825  public static final int 
DEFAULT_HBASE_META_BLOCK_SIZE = 8 * 1024;
+826
+827  /**
+828   * Parameter name for unique identifier 
for this {@link org.apache.hadoop.conf.Configuration}

[23/51] [partial] hbase-site git commit: Published site at .

2017-11-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html 
b/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html
index 741fc75..a0a7ebb 100644
--- a/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html
+++ b/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html
@@ -115,7 +115,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class TBoundedThreadPoolServer
+public class TBoundedThreadPoolServer
 extends org.apache.thrift.server.TServer
 A bounded thread pool server customized for HBase.
 
@@ -332,7 +332,7 @@ extends org.apache.thrift.server.TServer
 
 
 QUEUE_FULL_MSG
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String QUEUE_FULL_MSG
+private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String QUEUE_FULL_MSG
 
 See Also:
 Constant
 Field Values
@@ -345,7 +345,7 @@ extends org.apache.thrift.server.TServer
 
 
 MIN_WORKER_THREADS_CONF_KEY
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MIN_WORKER_THREADS_CONF_KEY
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MIN_WORKER_THREADS_CONF_KEY
 The "core size" of the thread pool. New threads are created 
on every
  connection until this many threads are created.
 
@@ -360,7 +360,7 @@ extends org.apache.thrift.server.TServer
 
 
 DEFAULT_MIN_WORKER_THREADS
-public static finalint DEFAULT_MIN_WORKER_THREADS
+public static finalint DEFAULT_MIN_WORKER_THREADS
 This default core pool size should be enough for many test 
scenarios. We
  want to override this with a much larger number (e.g. at least 200) for a
  large-scale production setup.
@@ -376,7 +376,7 @@ extends org.apache.thrift.server.TServer
 
 
 MAX_WORKER_THREADS_CONF_KEY
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MAX_WORKER_THREADS_CONF_KEY
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MAX_WORKER_THREADS_CONF_KEY
 The maximum size of the thread pool. When the pending 
request queue
  overflows, new threads are created until their number reaches this number.
  After that, the server starts dropping connections.
@@ -392,7 +392,7 @@ extends org.apache.thrift.server.TServer
 
 
 DEFAULT_MAX_WORKER_THREADS
-public static finalint DEFAULT_MAX_WORKER_THREADS
+public static finalint DEFAULT_MAX_WORKER_THREADS
 
 See Also:
 Constant
 Field Values
@@ -405,7 +405,7 @@ extends org.apache.thrift.server.TServer
 
 
 MAX_QUEUED_REQUESTS_CONF_KEY
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MAX_QUEUED_REQUESTS_CONF_KEY
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MAX_QUEUED_REQUESTS_CONF_KEY
 The maximum number of pending connections waiting in the 
queue. If there
  are no idle threads in the pool, the server queues requests. Only when
  the queue overflows, new threads are added, up to
@@ -422,7 +422,7 @@ extends org.apache.thrift.server.TServer
 
 
 DEFAULT_MAX_QUEUED_REQUESTS
-public static finalint DEFAULT_MAX_QUEUED_REQUESTS
+public static finalint DEFAULT_MAX_QUEUED_REQUESTS
 
 See Also:
 Constant
 Field Values
@@ -435,7 +435,7 @@ extends org.apache.thrift.server.TServer
 
 
 THREAD_KEEP_ALIVE_TIME_SEC_CONF_KEY
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String THREAD_KEEP_ALIVE_TIME_SEC_CONF_KEY
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String THREAD_KEEP_ALIVE_TIME_SEC_CONF_KEY
 Default amount of time in seconds to keep a thread alive. 
Worker threads
  are stopped after being idle for this long.
 
@@ -450,7 +450,7 @@ extends org.apache.thrift.server.TServer
 
 
 DEFAULT_THREAD_KEEP_ALIVE_TIME_SEC
-private static finalint DEFAULT_THREAD_KEEP_ALIVE_TIME_SEC
+private static finalint DEFAULT_THREAD_KEEP_ALIVE_TIME_SEC
 
 See Also:
 Constant
 Field Values
@@ -463,7 +463,7 @@ extends org.apache.thrift.server.TServer
 
 
 TIME_TO_WAIT_AFTER_SHUTDOWN_MS
-public static finalint TIME_TO_WAIT_AFTER_SHUTDOWN_MS
+public 

[23/51] [partial] hbase-site git commit: Published site at .

2017-11-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9722a17/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
index 40cd159..2da0903 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
@@ -260,7 +260,7 @@
 252
 253// Update meta events (for testing)
 254if (hasProcExecutor) {
-255  
getProcedureScheduler().suspendEvent(metaLoadEvent);
+255  metaLoadEvent.suspend();
 256  setFailoverCleanupDone(false);
 257  for (RegionInfo hri: 
getMetaRegionSet()) {
 258setMetaInitialized(hri, false);
@@ -421,1455 +421,1454 @@
 413  }
 414
 415  public boolean 
waitMetaInitialized(final Procedure proc, final RegionInfo regionInfo) {
-416return 
getProcedureScheduler().waitEvent(
-417  
getMetaInitializedEvent(getMetaForRegion(regionInfo)), proc);
-418  }
-419
-420  private void setMetaInitialized(final 
RegionInfo metaRegionInfo, final boolean isInitialized) {
-421assert isMetaRegion(metaRegionInfo) : 
"unexpected non-meta region " + metaRegionInfo;
-422final ProcedureEvent metaInitEvent = 
getMetaInitializedEvent(metaRegionInfo);
-423if (isInitialized) {
-424  
getProcedureScheduler().wakeEvent(metaInitEvent);
-425} else {
-426  
getProcedureScheduler().suspendEvent(metaInitEvent);
-427}
-428  }
-429
-430  private ProcedureEvent 
getMetaInitializedEvent(final RegionInfo metaRegionInfo) {
-431assert isMetaRegion(metaRegionInfo) : 
"unexpected non-meta region " + metaRegionInfo;
-432// TODO: handle multiple meta.
-433return metaInitializedEvent;
-434  }
-435
-436  public boolean waitMetaLoaded(final 
Procedure proc) {
-437return 
getProcedureScheduler().waitEvent(metaLoadEvent, proc);
-438  }
-439
-440  protected void wakeMetaLoadedEvent() 
{
-441
getProcedureScheduler().wakeEvent(metaLoadEvent);
-442assert isMetaLoaded() : "expected 
meta to be loaded";
-443  }
-444
-445  public boolean isMetaLoaded() {
-446return metaLoadEvent.isReady();
-447  }
-448
-449  // 

-450  //  TODO: Sync helpers
-451  // 

-452  public void assignMeta(final RegionInfo 
metaRegionInfo) throws IOException {
-453assignMeta(metaRegionInfo, null);
-454  }
-455
-456  public void assignMeta(final RegionInfo 
metaRegionInfo, final ServerName serverName)
-457  throws IOException {
-458assert isMetaRegion(metaRegionInfo) : 
"unexpected non-meta region " + metaRegionInfo;
-459AssignProcedure proc;
-460if (serverName != null) {
-461  LOG.debug("Try assigning Meta " + 
metaRegionInfo + " to " + serverName);
-462  proc = 
createAssignProcedure(metaRegionInfo, serverName);
-463} else {
-464  LOG.debug("Assigning " + 
metaRegionInfo.getRegionNameAsString());
-465  proc = 
createAssignProcedure(metaRegionInfo, false);
-466}
-467
ProcedureSyncWait.submitAndWaitProcedure(master.getMasterProcedureExecutor(), 
proc);
-468  }
-469
-470  /**
-471   * Start a new thread to check if there 
are region servers whose versions are higher than others.
-472   * If so, move all system table regions 
to RS with the highest version to keep compatibility.
-473   * The reason is, RS in new version may 
not be able to access RS in old version when there are
-474   * some incompatible changes.
-475   */
-476  public void 
checkIfShouldMoveSystemRegionAsync() {
-477new Thread(() - {
-478  try {
-479synchronized 
(checkIfShouldMoveSystemRegionLock) {
-480  ListRegionPlan plans = 
new ArrayList();
-481  for (ServerName server : 
getExcludedServersForSystemTable()) {
-482if 
(master.getServerManager().isServerDead(server)) {
-483  // TODO: See HBASE-18494 
and HBASE-18495. Though getExcludedServersForSystemTable()
-484  // considers only online 
servers, the server could be queued for dead server
-485  // processing. As region 
assignments for crashed server is handled by
-486  // ServerCrashProcedure, do 
NOT handle them here. The goal is to handle this through
-487  // regular flow of 
LoadBalancer as a favored node and not to have this special
-488  // handling.
-489  continue;
-490}
-491ListRegionInfo 
regionsShouldMove = getCarryingSystemTables(server);
-492  

[23/51] [partial] hbase-site git commit: Published site at .

2017-11-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b1eb7453/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html
index d941aa1..7b68289 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class HMobStore
+public class HMobStore
 extends HStore
 The store implementation to save MOBs (medium objects), it 
extends the HStore.
  When a descriptor of a column family has the value "IS_MOB", it means this 
column family
@@ -501,7 +501,7 @@ extends 
 
 LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static finalorg.apache.commons.logging.Log LOG
 
 
 
@@ -510,7 +510,7 @@ extends 
 
 mobCacheConfig
-privateMobCacheConfig mobCacheConfig
+privateMobCacheConfig mobCacheConfig
 
 
 
@@ -519,7 +519,7 @@ extends 
 
 homePath
-privateorg.apache.hadoop.fs.Path homePath
+privateorg.apache.hadoop.fs.Path homePath
 
 
 
@@ -528,7 +528,7 @@ extends 
 
 mobFamilyPath
-privateorg.apache.hadoop.fs.Path mobFamilyPath
+privateorg.apache.hadoop.fs.Path mobFamilyPath
 
 
 
@@ -537,7 +537,7 @@ extends 
 
 cellsCountCompactedToMob
-private volatilelong cellsCountCompactedToMob
+private volatilelong cellsCountCompactedToMob
 
 
 
@@ -546,7 +546,7 @@ extends 
 
 cellsCountCompactedFromMob
-private volatilelong cellsCountCompactedFromMob
+private volatilelong cellsCountCompactedFromMob
 
 
 
@@ -555,7 +555,7 @@ extends 
 
 cellsSizeCompactedToMob
-private volatilelong cellsSizeCompactedToMob
+private volatilelong cellsSizeCompactedToMob
 
 
 
@@ -564,7 +564,7 @@ extends 
 
 cellsSizeCompactedFromMob
-private volatilelong cellsSizeCompactedFromMob
+private volatilelong cellsSizeCompactedFromMob
 
 
 
@@ -573,7 +573,7 @@ extends 
 
 mobFlushCount
-private volatilelong mobFlushCount
+private volatilelong mobFlushCount
 
 
 
@@ -582,7 +582,7 @@ extends 
 
 mobFlushedCellsCount
-private volatilelong mobFlushedCellsCount
+private volatilelong mobFlushedCellsCount
 
 
 
@@ -591,7 +591,7 @@ extends 
 
 mobFlushedCellsSize
-private volatilelong mobFlushedCellsSize
+private volatilelong mobFlushedCellsSize
 
 
 
@@ -600,7 +600,7 @@ extends 
 
 mobScanCellsCount
-private volatilelong mobScanCellsCount
+private volatilelong mobScanCellsCount
 
 
 
@@ -609,7 +609,7 @@ extends 
 
 mobScanCellsSize
-private volatilelong mobScanCellsSize
+private volatilelong mobScanCellsSize
 
 
 
@@ -618,7 +618,7 @@ extends 
 
 family
-privateColumnFamilyDescriptor 
family
+privateColumnFamilyDescriptor 
family
 
 
 
@@ -627,7 +627,7 @@ extends 
 
 map
-privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.Path map
+privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.Path map
 
 
 
@@ -636,7 +636,7 @@ extends 
 
 keyLock
-private finalIdLock keyLock
+private finalIdLock keyLock
 
 
 
@@ -645,7 +645,7 @@ extends 
 
 refCellTags
-private finalbyte[] refCellTags
+private finalbyte[] refCellTags
 
 
 
@@ -662,7 +662,7 @@ extends 
 
 HMobStore
-publicHMobStore(HRegionregion,
+publicHMobStore(HRegionregion,
  ColumnFamilyDescriptorfamily,
  org.apache.hadoop.conf.ConfigurationconfParam)
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
@@ -686,7 +686,7 @@ extends 
 
 createCacheConf
-protectedvoidcreateCacheConf(ColumnFamilyDescriptorfamily)
+protectedvoidcreateCacheConf(ColumnFamilyDescriptorfamily)
 Creates the mob cache config.
 
 Overrides:
@@ -702,7 +702,7 @@ extends 
 
 getConfiguration
-publicorg.apache.hadoop.conf.ConfigurationgetConfiguration()
+publicorg.apache.hadoop.conf.ConfigurationgetConfiguration()
 Gets current config.
 
 
@@ -712,7 +712,7 @@ extends 
 
 createScanner
-protectedKeyValueScannercreateScanner(Scanscan,
+protectedKeyValueScannercreateScanner(Scanscan,
 ScanInfoscanInfo,
 

[23/51] [partial] hbase-site git commit: Published site at .

2017-11-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9118853f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html
index a9e8ff7..c670860 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html
@@ -65,1723 +65,1733 @@
 057import 
org.apache.hadoop.hbase.client.TableDescriptor;
 058import 
org.apache.hadoop.hbase.coprocessor.BaseEnvironment;
 059import 
org.apache.hadoop.hbase.coprocessor.BulkLoadObserver;
-060import 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
-061import 
org.apache.hadoop.hbase.coprocessor.CoprocessorService;
-062import 
org.apache.hadoop.hbase.coprocessor.CoprocessorServiceBackwardCompatiblity;
-063import 
org.apache.hadoop.hbase.coprocessor.CoreCoprocessor;
-064import 
org.apache.hadoop.hbase.coprocessor.EndpointObserver;
-065import 
org.apache.hadoop.hbase.coprocessor.HasRegionServerServices;
-066import 
org.apache.hadoop.hbase.coprocessor.MetricsCoprocessor;
-067import 
org.apache.hadoop.hbase.coprocessor.ObserverContext;
-068import 
org.apache.hadoop.hbase.coprocessor.RegionCoprocessor;
-069import 
org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
-070import 
org.apache.hadoop.hbase.coprocessor.RegionObserver;
-071import 
org.apache.hadoop.hbase.coprocessor.RegionObserver.MutationType;
-072import 
org.apache.hadoop.hbase.filter.ByteArrayComparable;
-073import 
org.apache.hadoop.hbase.io.FSDataInputStreamWrapper;
-074import 
org.apache.hadoop.hbase.io.Reference;
-075import 
org.apache.hadoop.hbase.io.hfile.CacheConfig;
-076import 
org.apache.hadoop.hbase.metrics.MetricRegistry;
-077import 
org.apache.hadoop.hbase.regionserver.Region.Operation;
-078import 
org.apache.hadoop.hbase.regionserver.compactions.CompactionLifeCycleTracker;
-079import 
org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
-080import 
org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker;
-081import 
org.apache.hadoop.hbase.security.User;
-082import 
org.apache.hadoop.hbase.util.Bytes;
-083import 
org.apache.hadoop.hbase.util.CoprocessorClassLoader;
-084import 
org.apache.hadoop.hbase.util.Pair;
-085import 
org.apache.hadoop.hbase.wal.WALEdit;
-086import 
org.apache.hadoop.hbase.wal.WALKey;
-087import 
org.apache.yetus.audience.InterfaceAudience;
-088
-089import 
org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;
-090
-091/**
-092 * Implements the coprocessor environment 
and runtime support for coprocessors
-093 * loaded within a {@link Region}.
-094 */
-095@InterfaceAudience.Private
-096public class RegionCoprocessorHost
-097extends 
CoprocessorHostRegionCoprocessor, RegionCoprocessorEnvironment {
-098
-099  private static final Log LOG = 
LogFactory.getLog(RegionCoprocessorHost.class);
-100  // The shared data map
-101  private static final 
ReferenceMapString, ConcurrentMapString, Object SHARED_DATA_MAP 
=
-102  new 
ReferenceMap(AbstractReferenceMap.ReferenceStrength.HARD,
-103  
AbstractReferenceMap.ReferenceStrength.WEAK);
-104
-105  // optimization: no need to call 
postScannerFilterRow, if no coprocessor implements it
-106  private final boolean 
hasCustomPostScannerFilterRow;
-107
-108  /**
-109   *
-110   * Encapsulation of the environment of 
each coprocessor
-111   */
-112  private static class RegionEnvironment 
extends BaseEnvironmentRegionCoprocessor
-113  implements 
RegionCoprocessorEnvironment {
-114private Region region;
-115ConcurrentMapString, Object 
sharedData;
-116private final MetricRegistry 
metricRegistry;
-117private final RegionServerServices 
services;
-118
-119/**
-120 * Constructor
-121 * @param impl the coprocessor 
instance
-122 * @param priority chaining 
priority
-123 */
-124public RegionEnvironment(final 
RegionCoprocessor impl, final int priority,
-125final int seq, final 
Configuration conf, final Region region,
-126final RegionServerServices 
services, final ConcurrentMapString, Object sharedData) {
-127  super(impl, priority, seq, conf);
-128  this.region = region;
-129  this.sharedData = sharedData;
-130  this.services = services;
-131  this.metricRegistry =
-132  
MetricsCoprocessor.createRegistryForRegionCoprocessor(impl.getClass().getName());
-133}
-134
-135/** @return the region */
-136@Override
-137public Region getRegion() {
-138  return region;
-139}
-140
-141public 

  1   2   3   >