[27/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.

2019-01-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html 
b/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html
index 9913946..055a934 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html
@@ -236,18 +236,22 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-org.apache.hadoop.hbase.tool
+org.apache.hadoop.hbase.thrift2.client
 
 
 
-org.apache.hadoop.hbase.util
+org.apache.hadoop.hbase.tool
 
 
 
-org.apache.hadoop.hbase.wal
+org.apache.hadoop.hbase.util
 
 
 
+org.apache.hadoop.hbase.wal
+
+
+
 org.apache.hadoop.hbase.zookeeper
 
 
@@ -1948,6 +1952,40 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
+
+
+
+Uses of Pair in org.apache.hadoop.hbase.thrift2.client
+
+Methods in org.apache.hadoop.hbase.thrift2.client
 that return Pair
+
+Modifier and Type
+Method and Description
+
+
+
+Pairhttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer
+ThriftAdmin.getAlterStatus(byte[]tableName)
+
+
+Pairhttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer
+ThriftAdmin.getAlterStatus(TableNametableName)
+
+
+Pairorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Client,org.apache.thrift.transport.TTransport
+ThriftConnection.DefaultThriftClientBuilder.getClient()
+
+
+Pairorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Client,org.apache.thrift.transport.TTransport
+ThriftConnection.HTTPThriftClientBuilder.getClient()
+
+
+abstract Pairorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Client,org.apache.thrift.transport.TTransport
+ThriftClientBuilder.getClient()
+
+
+
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
index 25ec8fe..4e1d4c5 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
@@ -541,14 +541,14 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.util.PrettyPrinter.Unit
+org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
+org.apache.hadoop.hbase.util.ChecksumType
 org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE
+org.apache.hadoop.hbase.util.PrettyPrinter.Unit
+org.apache.hadoop.hbase.util.PoolMap.PoolType
+org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer
 (implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
 org.apache.hadoop.hbase.util.Order
 org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer
 (implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
-org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer
 (implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
-org.apache.hadoop.hbase.util.PoolMap.PoolType
-org.apache.hadoop.hbase.util.ChecksumType
-org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/util/package-use.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-use.html 
b/devapidocs/org/apache/hadoop/hbase/util/package-use.html
index 6336a97..7c797bf 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/package-use.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/package-use.html
@@ -311,14 +311,18 @@ service.
 
 
 
-org.apache.hadoop.hbase.tool
+org.apache.hadoop.hbase.thrift2.client
 
 
 
-org.apache.hadoop.hbase.tool.coprocessor
+org.apache.hadoop.hbase.tool
 
 
 
+org.apache.hadoop.hbase.tool.coprocessor
+
+
+
 org.apache.hadoop.hbase.types
 
 
@@ -326,19 +330,19 @@ service.
  extensible data type API.
 

[27/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.

2019-01-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
index ff0e4d5..c3c0cb1 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
@@ -761,58 +761,65 @@
 
 
 default void
+MasterObserver.postIsRpcThrottleEnabled(ObserverContextMasterCoprocessorEnvironmentctx,
+booleanrpcThrottleEnabled)
+Called after getting if is rpc throttle enabled.
+
+
+
+default void
 MasterObserver.postListDecommissionedRegionServers(ObserverContextMasterCoprocessorEnvironmentctx)
 Called after list decommissioned region servers.
 
 
-
+
 default void
 MasterObserver.postListNamespaceDescriptors(ObserverContextMasterCoprocessorEnvironmentctx,
 https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListNamespaceDescriptordescriptors)
 Called after a listNamespaceDescriptors request has been 
processed.
 
 
-
+
 default void
 MasterObserver.postListReplicationPeers(ObserverContextMasterCoprocessorEnvironmentctx,
 https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringregex)
 Called after list replication peers.
 
 
-
+
 default void
 MasterObserver.postListRSGroups(ObserverContextMasterCoprocessorEnvironmentctx)
 Called after listing region server group information.
 
 
-
+
 default void
 MasterObserver.postListSnapshot(ObserverContextMasterCoprocessorEnvironmentctx,
 SnapshotDescriptionsnapshot)
 Called after listSnapshots request has been processed.
 
 
-
+
 default void
 MasterObserver.postLockHeartbeat(ObserverContextMasterCoprocessorEnvironmentctx)
 Called after heartbeat to a lock.
 
 
-
+
 default void
 RegionObserver.postMemStoreCompaction(ObserverContextRegionCoprocessorEnvironmentc,
   Storestore)
 Called after the in memory compaction is finished.
 
 
-
+
 default void
 MasterObserver.postMergeRegions(ObserverContextMasterCoprocessorEnvironmentc,
 RegionInfo[]regionsToMerge)
 called after merge regions request.
 
 
-
+
 default void
 MasterObserver.postMergeRegionsCommitAction(ObserverContextMasterCoprocessorEnvironmentctx,
 RegionInfo[]regionsToMerge,
@@ -820,7 +827,7 @@
 This will be called after META step as part of regions 
merge transaction.
 
 
-
+
 default void
 MasterObserver.postModifyNamespace(ObserverContextMasterCoprocessorEnvironmentctx,
NamespaceDescriptorcurrentNsDescriptor)
@@ -829,7 +836,7 @@
 
 
 
-
+
 default void
 MasterObserver.postModifyNamespace(ObserverContextMasterCoprocessorEnvironmentctx,
NamespaceDescriptoroldNsDescriptor,
@@ -837,7 +844,7 @@
 Called after the modifyNamespace operation has been 
requested.
 
 
-
+
 default void
 MasterObserver.postModifyTable(ObserverContextMasterCoprocessorEnvironmentctx,
TableNametableName,
@@ -847,7 +854,7 @@
 
 
 
-
+
 default void
 MasterObserver.postModifyTable(ObserverContextMasterCoprocessorEnvironmentctx,
TableNametableName,
@@ -856,7 +863,7 @@
 Called after the modifyTable operation has been 
requested.
 
 
-
+
 default void
 MasterObserver.postMove(ObserverContextMasterCoprocessorEnvironmentctx,
 RegionInforegion,
@@ -865,7 +872,7 @@
 Called after the region move has been requested.
 
 
-
+
 default void
 MasterObserver.postMoveServers(ObserverContextMasterCoprocessorEnvironmentctx,
https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in java.util">SetAddressservers,
@@ -873,7 +880,7 @@
 Called after servers are moved to target region server 
group
 
 
-
+
 default void
 MasterObserver.postMoveServersAndTables(ObserverContextMasterCoprocessorEnvironmentctx,
 https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in java.util">SetAddressservers,
@@ -882,7 +889,7 @@
 Called after servers are moved to target region server 
group
 
 
-
+
 default void
 MasterObserver.postMoveTables(ObserverContextMasterCoprocessorEnvironmentctx,
   https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in java.util">SetTableNametables,
@@ -890,7 +897,7 @@
 Called after servers are moved to target region server 
group
 
 
-
+
 default Cell
 RegionObserver.postMutationBeforeWAL(ObserverContextRegionCoprocessorEnvironmentctx,
  

[27/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.

2019-01-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
index cd0ff28..4f9947f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html
@@ -31,161 +31,161 @@
 023import static 
org.apache.hadoop.hbase.client.ConnectionUtils.resetController;
 024import static 
org.apache.hadoop.hbase.client.ConnectionUtils.translateException;
 025import static 
org.apache.hadoop.hbase.util.CollectionUtils.computeIfAbsent;
-026
-027import 
org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer;
-028
-029import java.io.IOException;
-030import java.util.ArrayList;
-031import java.util.Collections;
-032import java.util.HashMap;
-033import java.util.IdentityHashMap;
-034import java.util.List;
-035import java.util.Map;
-036import java.util.Optional;
-037import 
java.util.concurrent.CompletableFuture;
-038import 
java.util.concurrent.ConcurrentHashMap;
-039import 
java.util.concurrent.ConcurrentLinkedQueue;
-040import 
java.util.concurrent.ConcurrentMap;
-041import 
java.util.concurrent.ConcurrentSkipListMap;
-042import java.util.concurrent.TimeUnit;
-043import java.util.function.Supplier;
-044import java.util.stream.Collectors;
-045import java.util.stream.Stream;
-046
-047import 
org.apache.hadoop.hbase.CellScannable;
-048import 
org.apache.hadoop.hbase.DoNotRetryIOException;
-049import 
org.apache.hadoop.hbase.HRegionLocation;
-050import 
org.apache.hadoop.hbase.ServerName;
-051import 
org.apache.hadoop.hbase.TableName;
-052import 
org.apache.yetus.audience.InterfaceAudience;
-053import org.slf4j.Logger;
-054import org.slf4j.LoggerFactory;
-055import 
org.apache.hadoop.hbase.client.MultiResponse.RegionResult;
-056import 
org.apache.hadoop.hbase.client.RetriesExhaustedException.ThrowableWithExtraContext;
-057import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-058import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-059import 
org.apache.hadoop.hbase.shaded.protobuf.ResponseConverter;
-060import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-061import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService;
-062import 
org.apache.hadoop.hbase.util.Bytes;
-063import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-064
-065/**
-066 * Retry caller for batch.
-067 * p
-068 * Notice that, the {@link 
#operationTimeoutNs} is the total time limit now which is the same with
-069 * other single operations
-070 * p
-071 * And the {@link #maxAttempts} is a 
limit for each single operation in the batch logically. In the
-072 * implementation, we will record a 
{@code tries} parameter for each operation group, and if it is
-073 * split to several groups when retrying, 
the sub groups will inherit the {@code tries}. You can
-074 * imagine that the whole retrying 
process is a tree, and the {@link #maxAttempts} is the limit of
-075 * the depth of the tree.
-076 */
-077@InterfaceAudience.Private
-078class 
AsyncBatchRpcRetryingCallerT {
-079
-080  private static final Logger LOG = 
LoggerFactory.getLogger(AsyncBatchRpcRetryingCaller.class);
-081
-082  private final HashedWheelTimer 
retryTimer;
-083
-084  private final AsyncConnectionImpl 
conn;
-085
-086  private final TableName tableName;
-087
-088  private final ListAction 
actions;
-089
-090  private final 
ListCompletableFutureT futures;
-091
-092  private final 
IdentityHashMapAction, CompletableFutureT action2Future;
-093
-094  private final 
IdentityHashMapAction, ListThrowableWithExtraContext 
action2Errors;
-095
-096  private final long pauseNs;
-097
-098  private final int maxAttempts;
-099
-100  private final long 
operationTimeoutNs;
-101
-102  private final long rpcTimeoutNs;
-103
-104  private final int startLogErrorsCnt;
-105
-106  private final long startNs;
-107
-108  // we can not use HRegionLocation as 
the map key because the hashCode and equals method of
-109  // HRegionLocation only consider 
serverName.
-110  private static final class 
RegionRequest {
-111
-112public final HRegionLocation loc;
-113
-114public final 
ConcurrentLinkedQueueAction actions = new 
ConcurrentLinkedQueue();
-115
-116public RegionRequest(HRegionLocation 
loc) {
-117  this.loc = loc;
-118}
-119  }
-120
-121  private static final class 
ServerRequest {
-122
-123public final ConcurrentMapbyte[], 
RegionRequest actionsByRegion =
-124new 
ConcurrentSkipListMap(Bytes.BYTES_COMPARATOR);
-125
-126public void addAction(HRegionLocation 
loc, Action action) {
-127  computeIfAbsent(actionsByRegion, 

[27/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.

2019-01-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
index c4e8c8b..aa58108 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html
@@ -82,602 +82,613 @@
 074  public static final String 
USER_COPROCESSORS_ENABLED_CONF_KEY =
 075"hbase.coprocessor.user.enabled";
 076  public static final boolean 
DEFAULT_USER_COPROCESSORS_ENABLED = true;
-077
-078  private static final Logger LOG = 
LoggerFactory.getLogger(CoprocessorHost.class);
-079  protected Abortable abortable;
-080  /** Ordered set of loaded coprocessors 
with lock */
-081  protected final SortedListE 
coprocEnvironments =
-082  new SortedList(new 
EnvironmentPriorityComparator());
-083  protected Configuration conf;
-084  // unique file prefix to use for local 
copies of jars when classloading
-085  protected String pathPrefix;
-086  protected AtomicInteger loadSequence = 
new AtomicInteger();
-087
-088  public CoprocessorHost(Abortable 
abortable) {
-089this.abortable = abortable;
-090this.pathPrefix = 
UUID.randomUUID().toString();
-091  }
-092
-093  /**
-094   * Not to be confused with the 
per-object _coprocessors_ (above),
-095   * coprocessorNames is static and 
stores the set of all coprocessors ever
-096   * loaded by any thread in this JVM. It 
is strictly additive: coprocessors are
-097   * added to coprocessorNames, by 
checkAndLoadInstance() but are never removed, since
-098   * the intention is to preserve a 
history of all loaded coprocessors for
-099   * diagnosis in case of server crash 
(HBASE-4014).
-100   */
-101  private static SetString 
coprocessorNames =
-102  Collections.synchronizedSet(new 
HashSetString());
-103
-104  public static SetString 
getLoadedCoprocessors() {
-105synchronized (coprocessorNames) {
-106  return new 
HashSet(coprocessorNames);
-107}
-108  }
-109
-110  /**
-111   * Used to create a parameter to the 
HServerLoad constructor so that
-112   * HServerLoad can provide information 
about the coprocessors loaded by this
-113   * regionserver.
-114   * (HBASE-4070: Improve region server 
metrics to report loaded coprocessors
-115   * to master).
-116   */
-117  public SetString 
getCoprocessors() {
-118SetString returnValue = new 
TreeSet();
-119for (E e: coprocEnvironments) {
-120  
returnValue.add(e.getInstance().getClass().getSimpleName());
-121}
-122return returnValue;
-123  }
-124
-125  /**
-126   * Load system coprocessors once only. 
Read the class names from configuration.
-127   * Called by constructor.
-128   */
-129  protected void 
loadSystemCoprocessors(Configuration conf, String confKey) {
-130boolean coprocessorsEnabled = 
conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
-131  DEFAULT_COPROCESSORS_ENABLED);
-132if (!coprocessorsEnabled) {
-133  return;
-134}
-135
-136Class? implClass;
-137
-138// load default coprocessors from 
configure file
-139String[] defaultCPClasses = 
conf.getStrings(confKey);
-140if (defaultCPClasses == null || 
defaultCPClasses.length == 0)
-141  return;
-142
-143int priority = 
Coprocessor.PRIORITY_SYSTEM;
-144for (String className : 
defaultCPClasses) {
-145  className = className.trim();
-146  if (findCoprocessor(className) != 
null) {
-147// If already loaded will just 
continue
-148LOG.warn("Attempted duplicate 
loading of " + className + "; skipped");
-149continue;
-150  }
-151  ClassLoader cl = 
this.getClass().getClassLoader();
-152  
Thread.currentThread().setContextClassLoader(cl);
-153  try {
-154implClass = 
cl.loadClass(className);
-155// Add coprocessors as we go to 
guard against case where a coprocessor is specified twice
-156// in the configuration
-157E env = 
checkAndLoadInstance(implClass, priority, conf);
-158if (env != null) {
-159  
this.coprocEnvironments.add(env);
-160  LOG.info("System coprocessor {} 
loaded, priority={}.", className, priority);
-161  ++priority;
-162}
-163  } catch (Throwable t) {
-164// We always abort if system 
coprocessors cannot be loaded
-165abortServer(className, t);
-166  }
-167}
-168  }
-169
-170  /**
-171   * Load a coprocessor implementation 
into the host
-172   * @param path path to implementation 
jar
-173   * @param className the main class 
name
-174   * @param priority chaining priority
-175   * @param conf configuration for 
coprocessor

[27/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.

2019-01-01 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html
index ee11a41..79ad648 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html
@@ -120,6 +120,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
index 85b5b1b..ae5fa8d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html
@@ -191,6 +191,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
index 581a1bf..306c08d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html
@@ -120,6 +120,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
index dd01782..cebfb9d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
@@ -162,6 +162,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
index c6dda63..f1d19b2 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html
@@ -183,6 +183,6 @@
 
 
 
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072019 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.html
 

[27/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.

2018-12-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html
index 90e5a60..56edaab 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html
@@ -115,6 +115,13 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
+org.apache.hadoop.hbase.thrift2
+
+Provides an HBase http://thrift.apache.org/;>Thrift
+service.
+
+
+
 org.apache.hadoop.hbase.util
 
 
@@ -428,6 +435,37 @@ the order they are declared.
 
 
 
+
+
+
+Uses of BloomType in org.apache.hadoop.hbase.thrift2
+
+Methods in org.apache.hadoop.hbase.thrift2
 that return BloomType
+
+Modifier and Type
+Method and Description
+
+
+
+static BloomType
+ThriftUtilities.bloomFilterFromThrift(org.apache.hadoop.hbase.thrift2.generated.TBloomFilterTypein)
+
+
+
+
+Methods in org.apache.hadoop.hbase.thrift2
 with parameters of type BloomType
+
+Modifier and Type
+Method and Description
+
+
+
+static 
org.apache.hadoop.hbase.thrift2.generated.TBloomFilterType
+ThriftUtilities.bloomFilterFromHBase(BloomTypein)
+
+
+
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
index 3e892ef..9384f68 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
@@ -405,7 +405,7 @@
 HRegionServer.getRegionByEncodedName(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">StringencodedRegionName)
 
 
-(package private) static HRegion
+static HRegion
 HRegion.newHRegion(org.apache.hadoop.fs.PathtableDir,
   WALwal,
   org.apache.hadoop.fs.FileSystemfs,

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
index b06ddd9..bc92a6b 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html
@@ -606,7 +606,7 @@
 
 
 
-(package private) static HRegion
+static HRegion
 HRegion.newHRegion(org.apache.hadoop.fs.PathtableDir,
   WALwal,
   org.apache.hadoop.fs.FileSystemfs,

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index 8b58771..50c6f2d 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -716,20 +716,20 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
-org.apache.hadoop.hbase.regionserver.ScanType
-org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
-org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType
-org.apache.hadoop.hbase.regionserver.Region.Operation
-org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
 org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
 org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
-org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
+org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType
+org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type
 org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result

[27/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.

2018-12-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
index b8d0e70..ac5997f 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HRegionServer.CompactionChecker
+private static class HRegionServer.CompactionChecker
 extends ScheduledChore
 
 
@@ -233,7 +233,7 @@ extends 
 
 instance
-private finalHRegionServer instance
+private finalHRegionServer instance
 
 
 
@@ -242,7 +242,7 @@ extends 
 
 majorCompactPriority
-private finalint majorCompactPriority
+private finalint majorCompactPriority
 
 
 
@@ -251,7 +251,7 @@ extends 
 
 DEFAULT_PRIORITY
-private static finalint DEFAULT_PRIORITY
+private static finalint DEFAULT_PRIORITY
 
 See Also:
 Constant
 Field Values
@@ -264,7 +264,7 @@ extends 
 
 iteration
-privatelong iteration
+privatelong iteration
 
 
 
@@ -281,7 +281,7 @@ extends 
 
 CompactionChecker
-CompactionChecker(HRegionServerh,
+CompactionChecker(HRegionServerh,
   intsleepTime,
   Stoppablestopper)
 
@@ -300,7 +300,7 @@ extends 
 
 chore
-protectedvoidchore()
+protectedvoidchore()
 Description copied from 
class:ScheduledChore
 The task to execute on each scheduled execution of the 
Chore
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
index 5d4def3..30395a4 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HRegionServer.MovedRegionInfo
+private static class HRegionServer.MovedRegionInfo
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 
 
@@ -218,7 +218,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 serverName
-private finalServerName serverName
+private finalServerName serverName
 
 
 
@@ -227,7 +227,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 seqNum
-private finallong seqNum
+private finallong seqNum
 
 
 
@@ -236,7 +236,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 ts
-private finallong ts
+private finallong ts
 
 
 
@@ -253,7 +253,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 MovedRegionInfo
-publicMovedRegionInfo(ServerNameserverName,
+publicMovedRegionInfo(ServerNameserverName,
longcloseSeqNum)
 
 
@@ -271,7 +271,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getServerName
-publicServerNamegetServerName()
+publicServerNamegetServerName()
 
 
 
@@ -280,7 +280,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getSeqNum
-publiclonggetSeqNum()
+publiclonggetSeqNum()
 
 
 
@@ -289,7 +289,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getMoveTime
-publiclonggetMoveTime()
+publiclonggetMoveTime()
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
index 2a00361..fa924cd 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-protected static final class HRegionServer.MovedRegionsCleaner
+protected static final class HRegionServer.MovedRegionsCleaner
 extends ScheduledChore
 implements Stoppable
 Creates a Chore thread to clean the moved region 
cache.
@@ -242,7 +242,7 @@ implements 
 
 regionServer
-privateHRegionServer regionServer
+privateHRegionServer regionServer
 
 
 
@@ -251,7 +251,7 @@ implements 
 
 

[27/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.

2018-12-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
index ea05301..26a93dd 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
@@ -269,3590 +269,3574 @@
 261   */
 262  protected ClusterConnection 
clusterConnection;
 263
-264  /*
-265   * Long-living meta table locator, 
which is created when the server is started and stopped
-266   * when server shuts down. References 
to this locator shall be used to perform according
-267   * operations in EventHandlers. Primary 
reason for this decision is to make it mockable
-268   * for tests.
-269   */
-270  protected MetaTableLocator 
metaTableLocator;
-271
-272  /**
-273   * Go here to get table descriptors.
-274   */
-275  protected TableDescriptors 
tableDescriptors;
-276
-277  // Replication services. If no 
replication, this handler will be null.
-278  protected ReplicationSourceService 
replicationSourceHandler;
-279  protected ReplicationSinkService 
replicationSinkHandler;
-280
-281  // Compactions
-282  public CompactSplit 
compactSplitThread;
-283
-284  /**
-285   * Map of regions currently being 
served by this region server. Key is the
-286   * encoded region name.  All access 
should be synchronized.
-287   */
-288  protected final MapString, 
HRegion onlineRegions = new ConcurrentHashMap();
-289
-290  /**
-291   * Map of encoded region names to the 
DataNode locations they should be hosted on
-292   * We store the value as 
InetSocketAddress since this is used only in HDFS
-293   * API (create() that takes favored 
nodes as hints for placing file blocks).
-294   * We could have used ServerName here 
as the value class, but we'd need to
-295   * convert it to InetSocketAddress at 
some point before the HDFS API call, and
-296   * it seems a bit weird to store 
ServerName since ServerName refers to RegionServers
-297   * and here we really mean DataNode 
locations.
-298   */
-299  protected final MapString, 
InetSocketAddress[] regionFavoredNodesMap =
-300  new ConcurrentHashMap();
-301
-302  // Leases
-303  protected Leases leases;
+264  /**
+265   * Go here to get table descriptors.
+266   */
+267  protected TableDescriptors 
tableDescriptors;
+268
+269  // Replication services. If no 
replication, this handler will be null.
+270  protected ReplicationSourceService 
replicationSourceHandler;
+271  protected ReplicationSinkService 
replicationSinkHandler;
+272
+273  // Compactions
+274  public CompactSplit 
compactSplitThread;
+275
+276  /**
+277   * Map of regions currently being 
served by this region server. Key is the
+278   * encoded region name.  All access 
should be synchronized.
+279   */
+280  protected final MapString, 
HRegion onlineRegions = new ConcurrentHashMap();
+281
+282  /**
+283   * Map of encoded region names to the 
DataNode locations they should be hosted on
+284   * We store the value as 
InetSocketAddress since this is used only in HDFS
+285   * API (create() that takes favored 
nodes as hints for placing file blocks).
+286   * We could have used ServerName here 
as the value class, but we'd need to
+287   * convert it to InetSocketAddress at 
some point before the HDFS API call, and
+288   * it seems a bit weird to store 
ServerName since ServerName refers to RegionServers
+289   * and here we really mean DataNode 
locations.
+290   */
+291  protected final MapString, 
InetSocketAddress[] regionFavoredNodesMap =
+292  new ConcurrentHashMap();
+293
+294  // Leases
+295  protected Leases leases;
+296
+297  // Instance of the hbase executor 
executorService.
+298  protected ExecutorService 
executorService;
+299
+300  // If false, the file system has become 
unavailable
+301  protected volatile boolean fsOk;
+302  protected HFileSystem fs;
+303  protected HFileSystem walFs;
 304
-305  // Instance of the hbase executor 
executorService.
-306  protected ExecutorService 
executorService;
-307
-308  // If false, the file system has become 
unavailable
-309  protected volatile boolean fsOk;
-310  protected HFileSystem fs;
-311  protected HFileSystem walFs;
-312
-313  // Set when a report to the master 
comes back with a message asking us to
-314  // shutdown. Also set by call to stop 
when debugging or running unit tests
-315  // of HRegionServer in isolation.
-316  private volatile boolean stopped = 
false;
-317
-318  // Go down hard. Used if file system 
becomes unavailable and also in
-319  // debugging and unit tests.
-320  private volatile boolean 
abortRequested;
-321  public static final String 

[27/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.

2018-11-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html
deleted file mode 100644
index 908d0ce..000
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html
+++ /dev/null
@@ -1,165 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd;>
-
-
-
-
-
-Uses of Class 
org.apache.hadoop.hbase.master.assignment.ServerStateNode.ServerReportEvent 
(Apache HBase 3.0.0-SNAPSHOT API)
-
-
-
-
-
-
-
-JavaScript is disabled on your browser.
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-Prev
-Next
-
-
-Frames
-NoFrames
-
-
-AllClasses
-
-
-
-
-
-
-
-
-
-
-Uses of 
Classorg.apache.hadoop.hbase.master.assignment.ServerStateNode.ServerReportEvent
-
-
-
-
-
-Packages that use ServerStateNode.ServerReportEvent
-
-Package
-Description
-
-
-
-org.apache.hadoop.hbase.master.assignment
-
-
-
-
-
-
-
-
-
-
-Uses of ServerStateNode.ServerReportEvent
 in org.apache.hadoop.hbase.master.assignment
-
-Fields in org.apache.hadoop.hbase.master.assignment
 declared as ServerStateNode.ServerReportEvent
-
-Modifier and Type
-Field and Description
-
-
-
-private ServerStateNode.ServerReportEvent
-ServerStateNode.reportEvent
-
-
-
-
-
-
-
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-Prev
-Next
-
-
-Frames
-NoFrames
-
-
-AllClasses
-
-
-
-
-
-
-
-
-
-Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
-
-

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
index 44fe1fc..7e57352 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html
@@ -159,10 +159,6 @@
 
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCodestate,
 longseqId)
 
-
-protected void
-AssignmentManager.wakeServerReportEvent(ServerStateNodeserverNode)
-
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html
index 34c6cb8..03103e1 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html
@@ -37,7 +37,6 @@
 RegionStateStore
 RegionTransitionProcedure
 ServerStateNode
-ServerStateNode.ServerReportEvent
 SplitTableRegionProcedure
 TransitRegionStateProcedure
 UnassignProcedure

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
index 3235ced..577da39 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
@@ -210,22 +210,18 @@
 
 
 
-ServerStateNode.ServerReportEvent
-
-
-
 SplitTableRegionProcedure
 
 The procedure to split a region in a table.
 
 
-

[27/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.

2018-11-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html 
b/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html
index 44628e7..18eaf1f 100644
--- a/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html
+++ b/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html
@@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class RestoreSnapshotHelper
+public class RestoreSnapshotHelper
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 Helper to Restore/Clone a Snapshot
 
@@ -467,7 +467,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
@@ -476,7 +476,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 regionsMap
-private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Mapbyte[],byte[] regionsMap
+private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Mapbyte[],byte[] regionsMap
 
 
 
@@ -485,7 +485,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 parentsMap
-private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,Pairhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String parentsMap
+private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,Pairhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String parentsMap
 
 
 
@@ -494,7 +494,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 monitor
-private finalForeignExceptionDispatcher monitor
+private finalForeignExceptionDispatcher monitor
 
 
 
@@ -503,7 +503,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 status
-private finalMonitoredTask status
+private finalMonitoredTask status
 
 
 
@@ -512,7 +512,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 snapshotManifest
-private finalSnapshotManifest snapshotManifest
+private finalSnapshotManifest snapshotManifest
 
 
 
@@ -521,7 +521,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 snapshotDesc
-private 
finalorg.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription
 snapshotDesc
+private 
finalorg.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription
 snapshotDesc
 
 
 
@@ -530,7 +530,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 snapshotTable
-private finalTableName snapshotTable
+private finalTableName snapshotTable
 
 
 
@@ -539,7 +539,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 tableDesc
-private finalTableDescriptor tableDesc
+private finalTableDescriptor tableDesc
 
 
 
@@ -548,7 +548,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 rootDir
-private finalorg.apache.hadoop.fs.Path rootDir
+private finalorg.apache.hadoop.fs.Path rootDir
 
 
 
@@ -557,7 +557,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 tableDir
-private finalorg.apache.hadoop.fs.Path tableDir
+private finalorg.apache.hadoop.fs.Path tableDir
 
 
 
@@ -566,7 +566,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 conf
-private finalorg.apache.hadoop.conf.Configuration conf
+private finalorg.apache.hadoop.conf.Configuration conf
 
 
 
@@ -575,7 +575,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 fs
-private finalorg.apache.hadoop.fs.FileSystem fs
+private finalorg.apache.hadoop.fs.FileSystem fs
 
 
 
@@ -584,7 +584,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 

[27/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.

2018-10-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
index 0af8acd..c5f21ac 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html
@@ -645,1615 +645,1597 @@
 637
proc.afterReplay(getEnvironment());
 638  }
 639});
-640
-641// 4. Push the procedures to the 
timeout executor
-642waitingTimeoutList.forEach(proc - 
{
-643  
proc.afterReplay(getEnvironment());
-644  timeoutExecutor.add(proc);
-645});
-646// 5. restore locks
-647restoreLocks();
-648// 6. Push the procedure to the 
scheduler
-649
failedList.forEach(scheduler::addBack);
-650runnableList.forEach(p - {
-651  p.afterReplay(getEnvironment());
-652  if (!p.hasParent()) {
-653
sendProcedureLoadedNotification(p.getProcId());
-654  }
-655  // If the procedure holds the lock, 
put the procedure in front
-656  // If its parent holds the lock, 
put the procedure in front
-657  // TODO. Is that possible that its 
ancestor holds the lock?
-658  // For now, the deepest procedure 
hierarchy is:
-659  // ModifyTableProcedure - 
ReopenTableProcedure -
-660  // MoveTableProcedure - 
Unassign/AssignProcedure
-661  // But ModifyTableProcedure and 
ReopenTableProcedure won't hold the lock
-662  // So, check parent lock is 
enough(a tricky case is resovled by HBASE-21384).
-663  // If some one change or add new 
procedures making 'grandpa' procedure
-664  // holds the lock, but parent 
procedure don't hold the lock, there will
-665  // be a problem here. We have to 
check one procedure's ancestors.
-666  // And we need to change 
LockAndQueue.hasParentLock(Procedure? proc) method
-667  // to check all ancestors too.
-668  if (p.isLockedWhenLoading() || 
(p.hasParent()  procedures
-669  
.get(p.getParentProcId()).isLockedWhenLoading())) {
-670scheduler.addFront(p, false);
-671  } else {
-672// if it was not, it can wait.
-673scheduler.addBack(p, false);
-674  }
-675});
-676// After all procedures put into the 
queue, signal the worker threads.
-677// Otherwise, there is a race 
condition. See HBASE-21364.
-678scheduler.signalAll();
-679  }
+640// 4. restore locks
+641restoreLocks();
+642
+643// 5. Push the procedures to the 
timeout executor
+644waitingTimeoutList.forEach(proc - 
{
+645  
proc.afterReplay(getEnvironment());
+646  timeoutExecutor.add(proc);
+647});
+648
+649// 6. Push the procedure to the 
scheduler
+650
failedList.forEach(scheduler::addBack);
+651runnableList.forEach(p - {
+652  p.afterReplay(getEnvironment());
+653  if (!p.hasParent()) {
+654
sendProcedureLoadedNotification(p.getProcId());
+655  }
+656  scheduler.addBack(p);
+657});
+658// After all procedures put into the 
queue, signal the worker threads.
+659// Otherwise, there is a race 
condition. See HBASE-21364.
+660scheduler.signalAll();
+661  }
+662
+663  /**
+664   * Initialize the procedure executor, 
but do not start workers. We will start them later.
+665   * p/
+666   * It calls 
ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, 
and
+667   * ensure a single executor, and start 
the procedure replay to resume and recover the previous
+668   * pending and in-progress 
procedures.
+669   * @param numThreads number of threads 
available for procedure execution.
+670   * @param abortOnCorruption true if you 
want to abort your service in case a corrupted procedure
+671   *  is found on replay. 
otherwise false.
+672   */
+673  public void init(int numThreads, 
boolean abortOnCorruption) throws IOException {
+674// We have numThreads executor + one 
timer thread used for timing out
+675// procedures and triggering periodic 
procedures.
+676this.corePoolSize = numThreads;
+677this.maxPoolSize = 10 * numThreads;
+678LOG.info("Starting {} core workers 
(bigger of cpus/4 or 16) with max (burst) worker count={}",
+679corePoolSize, maxPoolSize);
 680
-681  /**
-682   * Initialize the procedure executor, 
but do not start workers. We will start them later.
-683   * p/
-684   * It calls 
ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, 
and
-685   * ensure a single executor, and start 
the procedure replay to resume and recover the previous
-686   * pending and in-progress 
procedures.
-687   * @param numThreads number of threads 
available for procedure 

[27/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.

2018-10-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
index c7d99b2..9d1542c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html
@@ -382,1357 +382,1365 @@
 374for (int i = 0; i  
this.curFunctionCosts.length; i++) {
 375  curFunctionCosts[i] = 
tempFunctionCosts[i];
 376}
-377LOG.info("start 
StochasticLoadBalancer.balancer, initCost=" + currentCost + ", functionCost="
-378+ functionCost());
+377double initCost = currentCost;
+378double newCost = currentCost;
 379
-380double initCost = currentCost;
-381double newCost = currentCost;
-382
-383long computedMaxSteps;
-384if (runMaxSteps) {
-385  computedMaxSteps = 
Math.max(this.maxSteps,
-386  ((long)cluster.numRegions * 
(long)this.stepsPerRegion * (long)cluster.numServers));
-387} else {
-388  computedMaxSteps = 
Math.min(this.maxSteps,
-389  ((long)cluster.numRegions * 
(long)this.stepsPerRegion * (long)cluster.numServers));
-390}
-391// Perform a stochastic walk to see 
if we can get a good fit.
-392long step;
-393
-394for (step = 0; step  
computedMaxSteps; step++) {
-395  Cluster.Action action = 
nextAction(cluster);
-396
-397  if (action.type == Type.NULL) {
-398continue;
-399  }
-400
-401  cluster.doAction(action);
-402  updateCostsWithAction(cluster, 
action);
-403
-404  newCost = computeCost(cluster, 
currentCost);
-405
-406  // Should this be kept?
-407  if (newCost  currentCost) {
-408currentCost = newCost;
-409
-410// save for JMX
-411curOverallCost = currentCost;
-412for (int i = 0; i  
this.curFunctionCosts.length; i++) {
-413  curFunctionCosts[i] = 
tempFunctionCosts[i];
-414}
-415  } else {
-416// Put things back the way they 
were before.
-417// TODO: undo by remembering old 
values
-418Action undoAction = 
action.undoAction();
-419cluster.doAction(undoAction);
-420updateCostsWithAction(cluster, 
undoAction);
-421  }
-422
-423  if 
(EnvironmentEdgeManager.currentTime() - startTime 
-424  maxRunningTime) {
-425break;
-426  }
-427}
-428long endTime = 
EnvironmentEdgeManager.currentTime();
-429
-430
metricsBalancer.balanceCluster(endTime - startTime);
-431
-432// update costs metrics
-433updateStochasticCosts(tableName, 
curOverallCost, curFunctionCosts);
-434if (initCost  currentCost) {
-435  plans = 
createRegionPlans(cluster);
-436  LOG.info("Finished computing new 
load balance plan. Computation took {}" +
-437" to try {} different iterations. 
 Found a solution that moves " +
-438"{} regions; Going from a 
computed cost of {}" +
-439" to a new cost of {}", 
java.time.Duration.ofMillis(endTime - startTime),
-440step, plans.size(), initCost, 
currentCost);
-441  return plans;
-442}
-443LOG.info("Could not find a better 
load balance plan.  Tried {} different configurations in " +
-444  "{}, and did not find anything with 
a computed cost less than {}", step,
-445  java.time.Duration.ofMillis(endTime 
- startTime), initCost);
-446return null;
-447  }
-448
-449  /**
-450   * update costs to JMX
-451   */
-452  private void 
updateStochasticCosts(TableName tableName, Double overall, Double[] subCosts) 
{
-453if (tableName == null) return;
-454
-455// check if the metricsBalancer is 
MetricsStochasticBalancer before casting
-456if (metricsBalancer instanceof 
MetricsStochasticBalancer) {
-457  MetricsStochasticBalancer balancer 
= (MetricsStochasticBalancer) metricsBalancer;
-458  // overall cost
-459  
balancer.updateStochasticCost(tableName.getNameAsString(),
-460"Overall", "Overall cost", 
overall);
-461
-462  // each cost function
-463  for (int i = 0; i  
costFunctions.length; i++) {
-464CostFunction costFunction = 
costFunctions[i];
-465String costFunctionName = 
costFunction.getClass().getSimpleName();
-466Double costPercent = (overall == 
0) ? 0 : (subCosts[i] / overall);
-467// TODO: cost function may need a 
specific description
-468
balancer.updateStochasticCost(tableName.getNameAsString(), costFunctionName,
-469  "The percent of " + 
costFunctionName, costPercent);
-470  }
-471}
-472  }
-473
-474  private String 

[27/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.

2018-10-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html
new file mode 100644
index 000..8ef0ff5
--- /dev/null
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html
@@ -0,0 +1,371 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/**
+002 * Licensed to the Apache Software 
Foundation (ASF) under one
+003 * or more contributor license 
agreements.  See the NOTICE file
+004 * distributed with this work for 
additional information
+005 * regarding copyright ownership.  The 
ASF licenses this file
+006 * to you under the Apache License, 
Version 2.0 (the
+007 * "License"); you may not use this file 
except in compliance
+008 * with the License.  You may obtain a 
copy of the License at
+009 *
+010 * 
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or 
agreed to in writing, software
+013 * distributed under the License is 
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+015 * See the License for the specific 
language governing permissions and
+016 * limitations under the License.
+017 */
+018package 
org.apache.hadoop.hbase.procedure2.store.wal;
+019
+020import java.io.IOException;
+021import java.util.ArrayList;
+022import java.util.Collection;
+023import java.util.Collections;
+024import java.util.Comparator;
+025import java.util.HashMap;
+026import java.util.Iterator;
+027import java.util.List;
+028import java.util.Map;
+029import 
java.util.NoSuchElementException;
+030import 
org.apache.commons.lang3.mutable.MutableInt;
+031import 
org.apache.hadoop.hbase.procedure2.Procedure;
+032import 
org.apache.hadoop.hbase.procedure2.ProcedureUtil;
+033import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
+034import 
org.apache.yetus.audience.InterfaceAudience;
+035import org.slf4j.Logger;
+036import org.slf4j.LoggerFactory;
+037
+038import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos;
+039
+040/**
+041 * Used to build the tree for 
procedures.
+042 * p/
+043 * We will group the procedures with the 
root procedure, and then validate each group. For each
+044 * group of procedures(with the same root 
procedure), we will collect all the stack ids, if the max
+045 * stack id is n, then all the stack ids 
should be from 0 to n, non-repetition and non-omission. If
+046 * not, we will consider all the 
procedures in this group as corrupted. Please see the code in
+047 * {@link #checkReady(Entry, Map)} 
method.
+048 * p/
+049 * For the procedures not in any group, 
i.e, can not find the root procedure for these procedures,
+050 * we will also consider them as 
corrupted. Please see the code in {@link #checkOrphan(Map)} method.
+051 */
+052@InterfaceAudience.Private
+053public final class WALProcedureTree {
+054
+055  private static final Logger LOG = 
LoggerFactory.getLogger(WALProcedureTree.class);
+056
+057  private static final class Entry {
+058
+059private final 
ProcedureProtos.Procedure proc;
+060
+061private final ListEntry 
subProcs = new ArrayList();
+062
+063public 
Entry(ProcedureProtos.Procedure proc) {
+064  this.proc = proc;
+065}
+066
+067@Override
+068public String toString() {
+069  StringBuilder sb = new 
StringBuilder();
+070  sb.append("Procedure(pid=");
+071  sb.append(proc.getProcId());
+072  sb.append(", ppid=");
+073  sb.append(proc.hasParentId() ? 
proc.getParentId() : Procedure.NO_PROC_ID);
+074  sb.append(", class=");
+075  sb.append(proc.getClassName());
+076  sb.append(")");
+077  return sb.toString();
+078}
+079  }
+080
+081  // when loading we will iterator the 
procedures twice, so use this class to cache the deserialized
+082  // result to prevent deserializing 
multiple times.
+083  private static final class ProtoAndProc 
{
+084private final 
ProcedureProtos.Procedure proto;
+085
+086private Procedure? proc;
+087
+088public 
ProtoAndProc(ProcedureProtos.Procedure proto) {
+089  this.proto = proto;
+090}
+091
+092public Procedure? getProc() 
throws IOException {
+093  if (proc == null) {
+094proc = 
ProcedureUtil.convertToProcedure(proto);
+095  }
+096  return proc;
+097}
+098  }
+099
+100  private final ListProtoAndProc 
validProcs = new ArrayList();
+101
+102  private final ListProtoAndProc 
corruptedProcs = new ArrayList();
+103
+104  private static boolean 
isFinished(ProcedureProtos.Procedure proc) {
+105if (!proc.hasParentId()) {
+106  switch (proc.getState()) 

[27/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.

2018-10-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
 
b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
index 3d27554..fd63f54 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-static class StochasticLoadBalancer.RegionReplicaCandidateGenerator
+static class StochasticLoadBalancer.RegionReplicaCandidateGenerator
 extends StochasticLoadBalancer.CandidateGenerator
 Generates candidates which moves the replicas out of the 
region server for
  co-hosted region replicas
@@ -228,7 +228,7 @@ extends 
 
 randomGenerator
-StochasticLoadBalancer.RandomCandidateGenerator
 randomGenerator
+StochasticLoadBalancer.RandomCandidateGenerator
 randomGenerator
 
 
 
@@ -245,7 +245,7 @@ extends 
 
 RegionReplicaCandidateGenerator
-RegionReplicaCandidateGenerator()
+RegionReplicaCandidateGenerator()
 
 
 
@@ -262,7 +262,7 @@ extends 
 
 selectCoHostedRegionPerGroup
-intselectCoHostedRegionPerGroup(int[]primariesOfRegionsPerGroup,
+intselectCoHostedRegionPerGroup(int[]primariesOfRegionsPerGroup,
  int[]regionsPerGroup,
  int[]regionIndexToPrimaryIndex)
 Randomly select one regionIndex out of all region replicas 
co-hosted in the same group
@@ -284,7 +284,7 @@ extends 
 
 generate
-BaseLoadBalancer.Cluster.Actiongenerate(BaseLoadBalancer.Clustercluster)
+BaseLoadBalancer.Cluster.Actiongenerate(BaseLoadBalancer.Clustercluster)
 
 Specified by:
 generatein
 classStochasticLoadBalancer.CandidateGenerator

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
 
b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
index 1faf326..76e8473 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-static class StochasticLoadBalancer.RegionReplicaHostCostFunction
+static class StochasticLoadBalancer.RegionReplicaHostCostFunction
 extends StochasticLoadBalancer.CostFunction
 A cost function for region replicas. We give a very high 
cost to hosting
  replicas of the same region in the same host. We do not prevent the case
@@ -272,7 +272,7 @@ extends 
 
 REGION_REPLICA_HOST_COST_KEY
-private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String REGION_REPLICA_HOST_COST_KEY
+private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String REGION_REPLICA_HOST_COST_KEY
 
 See Also:
 Constant
 Field Values
@@ -285,7 +285,7 @@ extends 
 
 DEFAULT_REGION_REPLICA_HOST_COST_KEY
-private static finalfloat DEFAULT_REGION_REPLICA_HOST_COST_KEY
+private static finalfloat DEFAULT_REGION_REPLICA_HOST_COST_KEY
 
 See Also:
 Constant
 Field Values
@@ -298,7 +298,7 @@ extends 
 
 maxCost
-long maxCost
+long maxCost
 
 
 
@@ -307,7 +307,7 @@ extends 
 
 costsPerGroup
-long[] costsPerGroup
+long[] costsPerGroup
 
 
 
@@ -316,7 +316,7 @@ extends 
 
 primariesOfRegionsPerGroup
-int[][] primariesOfRegionsPerGroup
+int[][] primariesOfRegionsPerGroup
 
 
 
@@ -333,7 +333,7 @@ extends 
 
 RegionReplicaHostCostFunction
-publicRegionReplicaHostCostFunction(org.apache.hadoop.conf.Configurationconf)
+publicRegionReplicaHostCostFunction(org.apache.hadoop.conf.Configurationconf)
 
 
 
@@ -350,7 +350,7 @@ extends 
 
 init
-voidinit(BaseLoadBalancer.Clustercluster)
+voidinit(BaseLoadBalancer.Clustercluster)
 Description copied from 
class:StochasticLoadBalancer.CostFunction
 Called once per LB invocation to give the cost function
  to initialize it's state, and perform any costly calculation.
@@ -366,7 +366,7 @@ extends 
 
 getMaxCost
-longgetMaxCost(BaseLoadBalancer.Clustercluster)
+longgetMaxCost(BaseLoadBalancer.Clustercluster)
 
 
 
@@ -375,7 +375,7 @@ extends 
 
 isNeeded
-booleanisNeeded()

[27/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.

2018-10-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html
--
diff --git a/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html 
b/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html
index 2d9b60f..6af42fe 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html
@@ -146,8 +146,8 @@
 
 java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.backup.TestBackupDeleteWithFailures.Failure
 org.apache.hadoop.hbase.backup.TestIncrementalBackupMergeWithFailures.FailurePhase
+org.apache.hadoop.hbase.backup.TestBackupDeleteWithFailures.Failure
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html 
b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
index 06d362a..cb399cd 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
@@ -1167,8 +1167,16 @@
 
 
 static HBaseClassTestRule
+TestRestoreSnapshotFromClientGetCompactionState.CLASS_RULE
+
+
+static HBaseClassTestRule
 TestMobCloneSnapshotFromClientNormal.CLASS_RULE
 
+
+static HBaseClassTestRule
+TestMobRestoreSnapshotFromClientClone.CLASS_RULE
+
 
 static HBaseClassTestRule
 TestCISleep.CLASS_RULE
@@ -1223,43 +1231,43 @@
 
 
 static HBaseClassTestRule
-TestAsyncTableGetMultiThreaded.CLASS_RULE
+TestMobRestoreSnapshotFromClientAfterSplittingRegions.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestSnapshotCloneIndependence.CLASS_RULE
+TestAsyncTableGetMultiThreaded.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestCIDeleteRpcTimeout.CLASS_RULE
+TestSnapshotCloneIndependence.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestZKAsyncRegistry.CLASS_RULE
+TestCIDeleteRpcTimeout.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestAsyncProcedureAdminApi.CLASS_RULE
+TestZKAsyncRegistry.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestAsyncClusterAdminApi2.CLASS_RULE
+TestAsyncProcedureAdminApi.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestCIGetOperationTimeout.CLASS_RULE
+TestAsyncClusterAdminApi2.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestAlwaysSetScannerId.CLASS_RULE
+TestCIGetOperationTimeout.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestSnapshotFromClient.CLASS_RULE
+TestAlwaysSetScannerId.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestRestoreSnapshotFromClient.CLASS_RULE
+TestSnapshotFromClient.CLASS_RULE
 
 
 static HBaseClassTestRule
@@ -1311,56 +1319,64 @@
 
 
 static HBaseClassTestRule
-TestBlockEvictionFromClient.CLASS_RULE
+TestMobRestoreSnapshotFromClientGetCompactionState.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestCIBadHostname.CLASS_RULE
+TestBlockEvictionFromClient.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE
+TestCIBadHostname.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestAsyncTableAdminApi2.CLASS_RULE
+TestCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestAsyncQuotaAdminApi.CLASS_RULE
+TestAsyncTableAdminApi2.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestAsyncMetaRegionLocator.CLASS_RULE
+TestAsyncQuotaAdminApi.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestAsyncTableNoncedRetry.CLASS_RULE
+TestAsyncMetaRegionLocator.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestRegionLocationCaching.CLASS_RULE
+TestAsyncTableNoncedRetry.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestConnectionUtils.CLASS_RULE
+TestRegionLocationCaching.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestMobCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE
+TestConnectionUtils.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestAsyncTableScanMetrics.CLASS_RULE
+TestMobCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE
 
 
 static HBaseClassTestRule
-TestFromClientSideScanExcpetionWithCoprocessor.CLASS_RULE
+TestAsyncTableScanMetrics.CLASS_RULE
 
 
 static HBaseClassTestRule
+TestFromClientSideScanExcpetionWithCoprocessor.CLASS_RULE
+
+
+static HBaseClassTestRule
 TestMobCloneSnapshotFromClientError.CLASS_RULE
 
+
+static HBaseClassTestRule
+TestRestoreSnapshotFromClientSchemaChange.CLASS_RULE
+
 
 static HBaseClassTestRule
 TestAsyncDecommissionAdminApi.CLASS_RULE
@@ 

[27/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.

2018-10-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
index 061ce80..bdfc3f8 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html
@@ -39,2126 +39,2163 @@
 031import java.util.Set;
 032import 
java.util.concurrent.ConcurrentHashMap;
 033import 
java.util.concurrent.CopyOnWriteArrayList;
-034import java.util.concurrent.TimeUnit;
-035import 
java.util.concurrent.atomic.AtomicBoolean;
-036import 
java.util.concurrent.atomic.AtomicInteger;
-037import 
java.util.concurrent.atomic.AtomicLong;
-038import java.util.stream.Collectors;
-039import java.util.stream.Stream;
-040import 
org.apache.hadoop.conf.Configuration;
-041import 
org.apache.hadoop.hbase.HConstants;
-042import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
-043import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-044import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
-045import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
-046import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
-047import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
-048import 
org.apache.hadoop.hbase.security.User;
-049import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-050import 
org.apache.hadoop.hbase.util.IdLock;
-051import 
org.apache.hadoop.hbase.util.NonceKey;
-052import 
org.apache.hadoop.hbase.util.Threads;
-053import 
org.apache.yetus.audience.InterfaceAudience;
-054import org.slf4j.Logger;
-055import org.slf4j.LoggerFactory;
-056
-057import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
-058import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+034import java.util.concurrent.Executor;
+035import java.util.concurrent.Executors;
+036import java.util.concurrent.TimeUnit;
+037import 
java.util.concurrent.atomic.AtomicBoolean;
+038import 
java.util.concurrent.atomic.AtomicInteger;
+039import 
java.util.concurrent.atomic.AtomicLong;
+040import java.util.stream.Collectors;
+041import java.util.stream.Stream;
+042import 
org.apache.hadoop.conf.Configuration;
+043import 
org.apache.hadoop.hbase.HConstants;
+044import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
+045import 
org.apache.hadoop.hbase.log.HBaseMarkers;
+046import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
+047import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
+048import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
+049import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureStoreListener;
+050import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
+051import 
org.apache.hadoop.hbase.security.User;
+052import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+053import 
org.apache.hadoop.hbase.util.IdLock;
+054import 
org.apache.hadoop.hbase.util.NonceKey;
+055import 
org.apache.hadoop.hbase.util.Threads;
+056import 
org.apache.yetus.audience.InterfaceAudience;
+057import org.slf4j.Logger;
+058import org.slf4j.LoggerFactory;
 059
-060import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
-061
-062/**
-063 * Thread Pool that executes the 
submitted procedures.
-064 * The executor has a ProcedureStore 
associated.
-065 * Each operation is logged and on 
restart the pending procedures are resumed.
-066 *
-067 * Unless the Procedure code throws an 
error (e.g. invalid user input)
-068 * the procedure will complete (at some 
point in time), On restart the pending
-069 * procedures are resumed and the once 
failed will be rolledback.
+060import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+061import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+062import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
+063
+064import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
+065
+066/**
+067 * Thread Pool that executes the 
submitted procedures.
+068 * The executor has a ProcedureStore 
associated.
+069 * Each operation is logged and on 
restart the pending procedures are resumed.
 070 *
-071 * The user can add procedures to the 
executor via submitProcedure(proc)
-072 * check for the finished state via 
isFinished(procId)
-073 * and get the result via 
getResult(procId)
-074 */
-075@InterfaceAudience.Private
-076public class 
ProcedureExecutorTEnvironment {

[27/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.

2018-09-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html
index c7f1e58..c74635e 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html
@@ -164,897 +164,923 @@
 156   */
 157  private volatile boolean bypass = 
false;
 158
-159  public boolean isBypass() {
-160return bypass;
-161  }
-162
-163  /**
-164   * set the bypass to true
-165   * Only called in {@link 
ProcedureExecutor#bypassProcedure(long, long, boolean)} for now,
-166   * DO NOT use this method alone, since 
we can't just bypass
-167   * one single procedure. We need to 
bypass its ancestor too. So making it package private
-168   */
-169  void bypass() {
-170this.bypass = true;
-171  }
+159  /**
+160   * Indicate whether we need to persist 
the procedure to ProcedureStore after execution. Default to
+161   * true, and the implementation can all 
{@link #skipPersistence()} to let the framework skip the
+162   * persistence of the procedure.
+163   * p/
+164   * This is useful when the procedure is 
in error and you want to retry later. The retry interval
+165   * and the number of retries are 
usually not critical so skip the persistence can save some
+166   * resources, and also speed up the 
restart processing.
+167   * p/
+168   * Notice that this value will be reset 
to true every time before execution. And when rolling back
+169   * we do not test this value.
+170   */
+171  private boolean persist = true;
 172
-173  /**
-174   * The main code of the procedure. It 
must be idempotent since execute()
-175   * may be called multiple times in case 
of machine failure in the middle
-176   * of the execution.
-177   * @param env the environment passed to 
the ProcedureExecutor
-178   * @return a set of sub-procedures to 
run or ourselves if there is more work to do or null if the
-179   * procedure is done.
-180   * @throws ProcedureYieldException the 
procedure will be added back to the queue and retried later.
-181   * @throws InterruptedException the 
procedure will be added back to the queue and retried later.
-182   * @throws ProcedureSuspendedException 
Signal to the executor that Procedure has suspended itself and
-183   * has set itself up waiting for an 
external event to wake it back up again.
-184   */
-185  protected abstract 
ProcedureTEnvironment[] execute(TEnvironment env)
-186throws ProcedureYieldException, 
ProcedureSuspendedException, InterruptedException;
-187
-188  /**
-189   * The code to undo what was done by 
the execute() code.
-190   * It is called when the procedure or 
one of the sub-procedures failed or an
-191   * abort was requested. It should 
cleanup all the resources created by
-192   * the execute() call. The 
implementation must be idempotent since rollback()
-193   * may be called multiple time in case 
of machine failure in the middle
-194   * of the execution.
-195   * @param env the environment passed to 
the ProcedureExecutor
-196   * @throws IOException temporary 
failure, the rollback will retry later
-197   * @throws InterruptedException the 
procedure will be added back to the queue and retried later
-198   */
-199  protected abstract void 
rollback(TEnvironment env)
-200throws IOException, 
InterruptedException;
-201
-202  /**
-203   * The abort() call is asynchronous and 
each procedure must decide how to deal
-204   * with it, if they want to be 
abortable. The simplest implementation
-205   * is to have an AtomicBoolean set in 
the abort() method and then the execute()
-206   * will check if the abort flag is set 
or not.
-207   * abort() may be called multiple times 
from the client, so the implementation
-208   * must be idempotent.
-209   *
-210   * pNOTE: abort() is not like 
Thread.interrupt(). It is just a notification
-211   * that allows the procedure 
implementor abort.
-212   */
-213  protected abstract boolean 
abort(TEnvironment env);
-214
-215  /**
-216   * The user-level code of the procedure 
may have some state to
-217   * persist (e.g. input arguments or 
current position in the processing state) to
-218   * be able to resume on failure.
-219   * @param serializer stores the 
serializable state
-220   */
-221  protected abstract void 
serializeStateData(ProcedureStateSerializer serializer)
-222throws IOException;
-223
-224  /**
-225   * Called on store load to allow the 
user to decode the previously serialized
-226   * state.
-227   * @param serializer contains the 
serialized state
-228   */
-229  protected abstract void 
deserializeStateData(ProcedureStateSerializer serializer)
-230throws IOException;
-231
-232  /**
-233   * The {@link 

[27/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.

2018-09-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
index 55d677f..aae23e1 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html
@@ -212,6 +212,10 @@
 HStore.getRegionFileSystem()
 
 
+(package private) HRegionFileSystem
+HRegion.getRegionWALFileSystem()
+
+
 static HRegionFileSystem
 HRegionFileSystem.openRegionFromFileSystem(org.apache.hadoop.conf.Configurationconf,
 org.apache.hadoop.fs.FileSystemfs,

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html
index 4c2b975..2abca97 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html
@@ -203,7 +203,7 @@
 static boolean
 WALSplitter.splitLogFile(org.apache.hadoop.fs.PathwalDir,
 org.apache.hadoop.fs.FileStatuslogfile,
-org.apache.hadoop.fs.FileSystemfs,
+org.apache.hadoop.fs.FileSystemwalFS,
 org.apache.hadoop.conf.Configurationconf,
 CancelableProgressablereporter,
 LastSequenceIdidChecker,
@@ -224,7 +224,7 @@
 WALSplitter(WALFactoryfactory,
org.apache.hadoop.conf.Configurationconf,
org.apache.hadoop.fs.PathwalDir,
-   org.apache.hadoop.fs.FileSystemfs,
+   org.apache.hadoop.fs.FileSystemwalFS,
LastSequenceIdidChecker,
SplitLogWorkerCoordinationsplitLogWorkerCoordination)
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
index 2ba4882..d9b1b13 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class AssignRegionHandler
+public class AssignRegionHandler
 extends EventHandler
 Handles opening of a region on a region server.
  
@@ -157,6 +157,10 @@ extends regionInfo
 
 
+private RetryCounter
+retryCounter
+
+
 private TableDescriptor
 tableDesc
 
@@ -266,7 +270,7 @@ extends 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
@@ -275,7 +279,7 @@ extends 
 
 regionInfo
-private finalRegionInfo regionInfo
+private finalRegionInfo regionInfo
 
 
 
@@ -284,16 +288,25 @@ extends 
 
 tableDesc
-private finalTableDescriptor tableDesc
+private finalTableDescriptor tableDesc
 
 
 
 
 
-
+
 
 masterSystemTime
-private finallong masterSystemTime
+private finallong masterSystemTime
+
+
+
+
+
+
+
+retryCounter
+private finalRetryCounter retryCounter
 
 
 
@@ -310,7 +323,7 @@ extends 
 
 AssignRegionHandler
-publicAssignRegionHandler(RegionServerServicesserver,
+publicAssignRegionHandler(RegionServerServicesserver,
RegionInforegionInfo,
@Nullable
TableDescriptortableDesc,
@@ -332,7 +345,7 @@ extends 
 
 getServer
-privateRegionServerServicesgetServer()
+privateRegionServerServicesgetServer()
 
 
 
@@ -341,7 +354,7 @@ extends 
 
 cleanUpAndReportFailure
-privatevoidcleanUpAndReportFailure(https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOExceptionerror)
+privatevoidcleanUpAndReportFailure(https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOExceptionerror)
   throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
 Throws:
@@ -355,7 +368,7 @@ extends 
 
 process
-publicvoidprocess()
+publicvoidprocess()
  throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in 

[27/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.

2018-09-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
index 49f081b..33c9cc0 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html
@@ -35,309 +35,328 @@
 027import 
java.util.concurrent.BlockingQueue;
 028import 
java.util.concurrent.ConcurrentHashMap;
 029import 
java.util.concurrent.ConcurrentMap;
-030import 
java.util.concurrent.LinkedBlockingQueue;
-031import 
java.util.concurrent.ThreadPoolExecutor;
-032import java.util.concurrent.TimeUnit;
-033import 
java.util.concurrent.atomic.AtomicLong;
-034
-035import 
org.apache.yetus.audience.InterfaceAudience;
-036import org.slf4j.Logger;
-037import org.slf4j.LoggerFactory;
-038import 
org.apache.hadoop.hbase.monitoring.ThreadMonitoring;
+030import java.util.concurrent.Executors;
+031import 
java.util.concurrent.LinkedBlockingQueue;
+032import 
java.util.concurrent.ThreadPoolExecutor;
+033import java.util.concurrent.TimeUnit;
+034import 
java.util.concurrent.atomic.AtomicLong;
+035import 
org.apache.hadoop.hbase.monitoring.ThreadMonitoring;
+036import 
org.apache.yetus.audience.InterfaceAudience;
+037import org.slf4j.Logger;
+038import org.slf4j.LoggerFactory;
 039
 040import 
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
 041import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
 042import 
org.apache.hbase.thirdparty.com.google.common.collect.Maps;
-043import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-044
-045/**
-046 * This is a generic executor service. 
This component abstracts a
-047 * threadpool, a queue to which {@link 
EventType}s can be submitted,
-048 * and a 
codeRunnable/code that handles the object that is added to the 
queue.
-049 *
-050 * pIn order to create a new 
service, create an instance of this class and
-051 * then do: 
codeinstance.startExecutorService("myService");/code.  When 
done
-052 * call {@link #shutdown()}.
-053 *
-054 * pIn order to use the service 
created above, call
-055 * {@link #submit(EventHandler)}.
-056 */
-057@InterfaceAudience.Private
-058public class ExecutorService {
-059  private static final Logger LOG = 
LoggerFactory.getLogger(ExecutorService.class);
-060
-061  // hold the all the executors created 
in a map addressable by their names
-062  private final 
ConcurrentHashMapString, Executor executorMap = new 
ConcurrentHashMap();
+043import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ListenableFuture;
+044import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ListeningScheduledExecutorService;
+045import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.MoreExecutors;
+046import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
+047
+048/**
+049 * This is a generic executor service. 
This component abstracts a
+050 * threadpool, a queue to which {@link 
EventType}s can be submitted,
+051 * and a 
codeRunnable/code that handles the object that is added to the 
queue.
+052 *
+053 * pIn order to create a new 
service, create an instance of this class and
+054 * then do: 
codeinstance.startExecutorService("myService");/code.  When 
done
+055 * call {@link #shutdown()}.
+056 *
+057 * pIn order to use the service 
created above, call
+058 * {@link #submit(EventHandler)}.
+059 */
+060@InterfaceAudience.Private
+061public class ExecutorService {
+062  private static final Logger LOG = 
LoggerFactory.getLogger(ExecutorService.class);
 063
-064  // Name of the server hosting this 
executor service.
-065  private final String servername;
+064  // hold the all the executors created 
in a map addressable by their names
+065  private final 
ConcurrentHashMapString, Executor executorMap = new 
ConcurrentHashMap();
 066
-067  /**
-068   * Default constructor.
-069   * @param servername Name of the 
hosting server.
-070   */
-071  public ExecutorService(final String 
servername) {
-072super();
-073this.servername = servername;
-074  }
-075
-076  /**
-077   * Start an executor service with a 
given name. If there was a service already
-078   * started with the same name, this 
throws a RuntimeException.
-079   * @param name Name of the service to 
start.
-080   */
-081  @VisibleForTesting
-082  public void startExecutorService(String 
name, int maxThreads) {
-083if (this.executorMap.get(name) != 
null) {
-084  throw new RuntimeException("An 
executor service with the name " + name +
-085" is already running!");
-086   

[27/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.

2018-09-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
index a5789e0..93a57cb 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html
@@ -238,4120 +238,4119 @@
 230 * @see Admin
 231 */
 232@InterfaceAudience.Private
-233@InterfaceStability.Evolving
-234public class HBaseAdmin implements Admin 
{
-235  private static final Logger LOG = 
LoggerFactory.getLogger(HBaseAdmin.class);
-236
-237  private ClusterConnection connection;
-238
-239  private final Configuration conf;
-240  private final long pause;
-241  private final int numRetries;
-242  private final int syncWaitTimeout;
-243  private boolean aborted;
-244  private int operationTimeout;
-245  private int rpcTimeout;
-246
-247  private RpcRetryingCallerFactory 
rpcCallerFactory;
-248  private RpcControllerFactory 
rpcControllerFactory;
-249
-250  private NonceGenerator ng;
-251
-252  @Override
-253  public int getOperationTimeout() {
-254return operationTimeout;
-255  }
-256
-257  HBaseAdmin(ClusterConnection 
connection) throws IOException {
-258this.conf = 
connection.getConfiguration();
-259this.connection = connection;
-260
-261// TODO: receive 
ConnectionConfiguration here rather than re-parsing these configs every time.
-262this.pause = 
this.conf.getLong(HConstants.HBASE_CLIENT_PAUSE,
-263
HConstants.DEFAULT_HBASE_CLIENT_PAUSE);
-264this.numRetries = 
this.conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER,
-265
HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER);
-266this.operationTimeout = 
this.conf.getInt(HConstants.HBASE_CLIENT_OPERATION_TIMEOUT,
-267
HConstants.DEFAULT_HBASE_CLIENT_OPERATION_TIMEOUT);
-268this.rpcTimeout = 
this.conf.getInt(HConstants.HBASE_RPC_TIMEOUT_KEY,
-269
HConstants.DEFAULT_HBASE_RPC_TIMEOUT);
-270this.syncWaitTimeout = 
this.conf.getInt(
-271  
"hbase.client.sync.wait.timeout.msec", 10 * 6); // 10min
-272
-273this.rpcCallerFactory = 
connection.getRpcRetryingCallerFactory();
-274this.rpcControllerFactory = 
connection.getRpcControllerFactory();
-275
-276this.ng = 
this.connection.getNonceGenerator();
-277  }
-278
-279  @Override
-280  public void abort(String why, Throwable 
e) {
-281// Currently does nothing but throw 
the passed message and exception
-282this.aborted = true;
-283throw new RuntimeException(why, e);
-284  }
-285
-286  @Override
-287  public boolean isAborted() {
-288return this.aborted;
-289  }
-290
-291  @Override
-292  public boolean abortProcedure(final 
long procId, final boolean mayInterruptIfRunning)
-293  throws IOException {
-294return 
get(abortProcedureAsync(procId, mayInterruptIfRunning), this.syncWaitTimeout,
-295  TimeUnit.MILLISECONDS);
-296  }
-297
-298  @Override
-299  public FutureBoolean 
abortProcedureAsync(final long procId, final boolean mayInterruptIfRunning)
-300  throws IOException {
-301Boolean abortProcResponse =
-302executeCallable(new 
MasterCallableAbortProcedureResponse(getConnection(),
-303getRpcControllerFactory()) 
{
-304  @Override
-305  protected AbortProcedureResponse 
rpcCall() throws Exception {
-306AbortProcedureRequest 
abortProcRequest =
-307
AbortProcedureRequest.newBuilder().setProcId(procId).build();
-308return 
master.abortProcedure(getRpcController(), abortProcRequest);
-309  }
-310}).getIsProcedureAborted();
-311return new AbortProcedureFuture(this, 
procId, abortProcResponse);
-312  }
-313
-314  @Override
-315  public ListTableDescriptor 
listTableDescriptors() throws IOException {
-316return 
listTableDescriptors((Pattern)null, false);
-317  }
-318
-319  @Override
-320  public ListTableDescriptor 
listTableDescriptors(Pattern pattern) throws IOException {
-321return listTableDescriptors(pattern, 
false);
-322  }
-323
-324  @Override
-325  public ListTableDescriptor 
listTableDescriptors(Pattern pattern, boolean includeSysTables)
-326  throws IOException {
-327return executeCallable(new 
MasterCallableListTableDescriptor(getConnection(),
-328getRpcControllerFactory()) {
-329  @Override
-330  protected 
ListTableDescriptor rpcCall() throws Exception {
-331GetTableDescriptorsRequest req 
=
-332
RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables);
-333return 
ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(),
-334req));
-335  }
-336});
-337  }
-338
-339  @Override
-340  

[27/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.

2018-09-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
index 2036bc9..bcfe1b9 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
@@ -28,113 +28,124 @@
 020
 021import java.io.IOException;
 022import java.util.ArrayList;
-023
-024import org.apache.hadoop.hbase.Cell;
-025import 
org.apache.yetus.audience.InterfaceAudience;
-026import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-027import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-028
-029import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
-030import 
org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException;
-031
-032/**
-033 * A filter that will only return the 
first KV from each row.
-034 * p
-035 * This filter can be used to more 
efficiently perform row count operations.
-036 */
-037@InterfaceAudience.Public
-038public class FirstKeyOnlyFilter extends 
FilterBase {
-039  private boolean foundKV = false;
-040
-041  public FirstKeyOnlyFilter() {
-042  }
-043
-044  @Override
-045  public void reset() {
-046foundKV = false;
-047  }
-048
-049  @Override
-050  public boolean filterRowKey(Cell cell) 
throws IOException {
-051// Impl in FilterBase might do 
unnecessary copy for Off heap backed Cells.
-052return false;
-053  }
-054
-055  @Deprecated
-056  @Override
-057  public ReturnCode filterKeyValue(final 
Cell c) {
-058return filterCell(c);
-059  }
-060
-061  @Override
-062  public ReturnCode filterCell(final Cell 
c) {
-063if(foundKV) return 
ReturnCode.NEXT_ROW;
-064foundKV = true;
-065return ReturnCode.INCLUDE;
-066  }
-067
-068  public static Filter 
createFilterFromArguments(ArrayListbyte [] filterArguments) {
-069
Preconditions.checkArgument(filterArguments.isEmpty(),
-070"Expected 
0 but got: %s", filterArguments.size());
-071return new FirstKeyOnlyFilter();
-072  }
-073
-074  /**
-075   * @return true if first KV has been 
found.
-076   */
-077  protected boolean hasFoundKV() {
-078return this.foundKV;
-079  }
-080
-081  /**
-082   *
-083   * @param value update {@link #foundKV} 
flag with value.
-084   */
-085  protected void setFoundKV(boolean 
value) {
-086this.foundKV = value;
-087  }
-088
-089  /**
-090   * @return The filter serialized using 
pb
-091   */
-092  @Override
-093  public byte [] toByteArray() {
-094
FilterProtos.FirstKeyOnlyFilter.Builder builder =
-095  
FilterProtos.FirstKeyOnlyFilter.newBuilder();
-096return 
builder.build().toByteArray();
-097  }
-098
-099  /**
-100   * @param pbBytes A pb serialized 
{@link FirstKeyOnlyFilter} instance
-101   * @return An instance of {@link 
FirstKeyOnlyFilter} made from codebytes/code
-102   * @throws 
org.apache.hadoop.hbase.exceptions.DeserializationException
-103   * @see #toByteArray
-104   */
-105  public static FirstKeyOnlyFilter 
parseFrom(final byte [] pbBytes)
-106  throws DeserializationException {
-107// There is nothing to deserialize.  
Why do this at all?
-108try {
-109  
FilterProtos.FirstKeyOnlyFilter.parseFrom(pbBytes);
-110} catch 
(InvalidProtocolBufferException e) {
-111  throw new 
DeserializationException(e);
-112}
-113// Just return a new instance.
-114return new FirstKeyOnlyFilter();
-115  }
-116
-117  /**
-118   * @param o the other filter to compare 
with
-119   * @return true if and only if the 
fields of the filter that are serialized
-120   * are equal to the corresponding 
fields in other.  Used for testing.
-121   */
-122  @Override
-123  boolean areSerializedFieldsEqual(Filter 
o) {
-124if (o == this) return true;
-125if (!(o instanceof 
FirstKeyOnlyFilter)) return false;
-126
-127return true;
-128  }
-129}
+023import java.util.Objects;
+024
+025import org.apache.hadoop.hbase.Cell;
+026import 
org.apache.yetus.audience.InterfaceAudience;
+027import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
+028import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
+029
+030import 
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+031import 
org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException;
+032
+033/**
+034 * A filter that will only return the 
first KV from each row.
+035 * p
+036 * This filter can be used to more 
efficiently perform row count operations.
+037 */
+038@InterfaceAudience.Public
+039public class FirstKeyOnlyFilter extends 
FilterBase {
+040  private boolean foundKV = false;
+041
+042  public FirstKeyOnlyFilter() {
+043  }
+044
+045  

[27/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.

2018-08-31 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
index 71b9375..8f7d034 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private final class HStore.StoreFlusherImpl
+private final class HStore.StoreFlusherImpl
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements StoreFlushContext
 
@@ -279,7 +279,7 @@ implements 
 
 tracker
-private finalFlushLifeCycleTracker tracker
+private finalFlushLifeCycleTracker tracker
 
 
 
@@ -288,7 +288,7 @@ implements 
 
 cacheFlushSeqNum
-private finallong cacheFlushSeqNum
+private finallong cacheFlushSeqNum
 
 
 
@@ -297,7 +297,7 @@ implements 
 
 snapshot
-privateMemStoreSnapshot snapshot
+privateMemStoreSnapshot snapshot
 
 
 
@@ -306,7 +306,7 @@ implements 
 
 tempFiles
-privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.Path tempFiles
+privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.Path tempFiles
 
 
 
@@ -315,7 +315,7 @@ implements 
 
 committedFiles
-privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.Path committedFiles
+privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.Path committedFiles
 
 
 
@@ -324,7 +324,7 @@ implements 
 
 cacheFlushCount
-privatelong cacheFlushCount
+privatelong cacheFlushCount
 
 
 
@@ -333,7 +333,7 @@ implements 
 
 cacheFlushSize
-privatelong cacheFlushSize
+privatelong cacheFlushSize
 
 
 
@@ -342,7 +342,7 @@ implements 
 
 outputFileSize
-privatelong outputFileSize
+privatelong outputFileSize
 
 
 
@@ -359,7 +359,7 @@ implements 
 
 StoreFlusherImpl
-privateStoreFlusherImpl(longcacheFlushSeqNum,
+privateStoreFlusherImpl(longcacheFlushSeqNum,
  FlushLifeCycleTrackertracker)
 
 
@@ -377,7 +377,7 @@ implements 
 
 prepare
-publicMemStoreSizeprepare()
+publicMemStoreSizeprepare()
 This is not thread safe. The caller should have a lock on 
the region or the store.
  If necessary, the lock can be added with the patch provided in 
HBASE-10087
 
@@ -394,7 +394,7 @@ implements 
 
 flushCache
-publicvoidflushCache(MonitoredTaskstatus)
+publicvoidflushCache(MonitoredTaskstatus)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Description copied from 
interface:StoreFlushContext
 Flush the cache (create the new store file)
@@ -415,7 +415,7 @@ implements 
 
 commit
-publicbooleancommit(MonitoredTaskstatus)
+publicbooleancommit(MonitoredTaskstatus)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Description copied from 
interface:StoreFlushContext
 Commit the flush - add the store file to the store and 
clear the
@@ -440,7 +440,7 @@ implements 
 
 getOutputFileSize
-publiclonggetOutputFileSize()
+publiclonggetOutputFileSize()
 
 Specified by:
 getOutputFileSizein
 interfaceStoreFlushContext
@@ -455,7 +455,7 @@ implements 
 
 getCommittedFiles
-publichttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.PathgetCommittedFiles()
+publichttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.fs.PathgetCommittedFiles()
 Description copied from 
interface:StoreFlushContext
 Returns the newly committed files from the flush. Called 
only if commit returns true
 
@@ -472,7 +472,7 @@ implements 
 
 replayFlush
-publicvoidreplayFlush(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringfileNames,
+publicvoidreplayFlush(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 

[27/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.

2018-08-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
index e42e654..fe4e081 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
@@ -560,7 +560,7 @@
 552TransitRegionStateProcedure proc;
 553regionNode.lock();
 554try {
-555  preTransitCheck(regionNode, 
RegionStates.STATES_EXPECTED_ON_OPEN);
+555  preTransitCheck(regionNode, 
STATES_EXPECTED_ON_ASSIGN);
 556  proc = 
TransitRegionStateProcedure.assign(getProcedureEnvironment(), regionInfo, 
sn);
 557  regionNode.setProcedure(proc);
 558} finally {
@@ -581,7 +581,7 @@
 573TransitRegionStateProcedure proc;
 574regionNode.lock();
 575try {
-576  preTransitCheck(regionNode, 
RegionStates.STATES_EXPECTED_ON_CLOSE);
+576  preTransitCheck(regionNode, 
STATES_EXPECTED_ON_UNASSIGN_OR_MOVE);
 577  proc = 
TransitRegionStateProcedure.unassign(getProcedureEnvironment(), regionInfo);
 578  regionNode.setProcedure(proc);
 579} finally {
@@ -599,7 +599,7 @@
 591TransitRegionStateProcedure proc;
 592regionNode.lock();
 593try {
-594  preTransitCheck(regionNode, 
RegionStates.STATES_EXPECTED_ON_CLOSE);
+594  preTransitCheck(regionNode, 
STATES_EXPECTED_ON_UNASSIGN_OR_MOVE);
 595  regionNode.checkOnline();
 596  proc = 
TransitRegionStateProcedure.move(getProcedureEnvironment(), regionInfo, 
targetServer);
 597  regionNode.setProcedure(proc);
@@ -1419,448 +1419,483 @@
 1411  }
 1412
 1413  // 

-1414  //  Region Status update
-1415  //  Should only be called in 
TransitRegionStateProcedure
-1416  // 

-1417  private void 
transitStateAndUpdate(RegionStateNode regionNode, RegionState.State newState,
-1418  RegionState.State... 
expectedStates) throws IOException {
-1419RegionState.State state = 
regionNode.getState();
-1420regionNode.transitionState(newState, 
expectedStates);
-1421boolean succ = false;
-1422try {
-1423  
regionStateStore.updateRegionLocation(regionNode);
-1424  succ = true;
-1425} finally {
-1426  if (!succ) {
-1427// revert
-1428regionNode.setState(state);
-1429  }
-1430}
-1431  }
-1432
-1433  // should be called within the 
synchronized block of RegionStateNode
-1434  void regionOpening(RegionStateNode 
regionNode) throws IOException {
-1435transitStateAndUpdate(regionNode, 
State.OPENING, RegionStates.STATES_EXPECTED_ON_OPEN);
-1436
regionStates.addRegionToServer(regionNode);
-1437// update the operation count 
metrics
-1438
metrics.incrementOperationCounter();
-1439  }
-1440
-1441  // should be called within the 
synchronized block of RegionStateNode.
-1442  // The parameter 'giveUp' means 
whether we will try to open the region again, if it is true, then
-1443  // we will persist the FAILED_OPEN 
state into hbase:meta.
-1444  void regionFailedOpen(RegionStateNode 
regionNode, boolean giveUp) throws IOException {
-1445RegionState.State state = 
regionNode.getState();
-1446ServerName regionLocation = 
regionNode.getRegionLocation();
-1447if (giveUp) {
-1448  
regionNode.setState(State.FAILED_OPEN);
-1449  
regionNode.setRegionLocation(null);
-1450  boolean succ = false;
-1451  try {
-1452
regionStateStore.updateRegionLocation(regionNode);
-1453succ = true;
-1454  } finally {
-1455if (!succ) {
-1456  // revert
-1457  regionNode.setState(state);
-1458  
regionNode.setRegionLocation(regionLocation);
-1459}
-1460  }
-1461}
-1462if (regionLocation != null) {
-1463  
regionStates.removeRegionFromServer(regionLocation, regionNode);
-1464}
-1465  }
-1466
-1467  // should be called within the 
synchronized block of RegionStateNode
-1468  void regionOpened(RegionStateNode 
regionNode) throws IOException {
-1469// TODO: OPENING Updates hbase:meta 
too... we need to do both here and there?
-1470// That is a lot of hbase:meta 
writing.
-1471transitStateAndUpdate(regionNode, 
State.OPEN, RegionStates.STATES_EXPECTED_ON_OPEN);
-1472RegionInfo hri = 
regionNode.getRegionInfo();
-1473if (isMetaRegion(hri)) {
-1474  // Usually we'd set a table 
ENABLED at this stage but hbase:meta is ALWAYs enabled, it
-1475  // can't be disabled -- so skip 
the RPC (besides... enabled is managed by TableStateManager

[27/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.

2018-08-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
index 2db4997..14fcebb 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
@@ -29,146 +29,161 @@
 021
 022import java.io.IOException;
 023import java.util.ArrayList;
-024
-025import org.apache.hadoop.hbase.Cell;
-026import 
org.apache.hadoop.hbase.CompareOperator;
-027import 
org.apache.yetus.audience.InterfaceAudience;
-028import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-029import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-030import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-031import 
org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException;
-032
-033/**
-034 * This filter is used to filter based on 
the key. It takes an operator
-035 * (equal, greater, not equal, etc) and a 
byte [] comparator for the row,
-036 * and column qualifier portions of a 
key.
-037 * p
-038 * This filter can be wrapped with {@link 
WhileMatchFilter} to add more control.
-039 * p
-040 * Multiple filters can be combined using 
{@link FilterList}.
-041 * p
-042 * If an already known row range needs to 
be scanned, 
-043 * use {@link 
org.apache.hadoop.hbase.CellScanner} start
-044 * and stop rows directly rather than a 
filter.
-045 */
-046@InterfaceAudience.Public
-047public class RowFilter extends 
CompareFilter {
-048
-049  private boolean filterOutRow = false;
-050
-051  /**
-052   * Constructor.
-053   * @param rowCompareOp the compare op 
for row matching
-054   * @param rowComparator the comparator 
for row matching
-055   * @deprecated Since 2.0.0. Will remove 
in 3.0.0. Use
-056   * {@link #RowFilter(CompareOperator, 
ByteArrayComparable)}} instead.
-057   */
-058  @Deprecated
-059  public RowFilter(final CompareOp 
rowCompareOp,
-060  final ByteArrayComparable 
rowComparator) {
-061super(rowCompareOp, rowComparator);
-062  }
-063
-064  /**
-065   * Constructor.
-066   * @param op the compare op for row 
matching
-067   * @param rowComparator the comparator 
for row matching
-068   */
-069  public RowFilter(final CompareOperator 
op,
-070   final 
ByteArrayComparable rowComparator) {
-071super(op, rowComparator);
-072  }
-073
-074  @Override
-075  public void reset() {
-076this.filterOutRow = false;
-077  }
-078
-079  @Deprecated
-080  @Override
-081  public ReturnCode filterKeyValue(final 
Cell c) {
-082return filterCell(c);
-083  }
-084
-085  @Override
-086  public ReturnCode filterCell(final Cell 
v) {
-087if(this.filterOutRow) {
-088  return ReturnCode.NEXT_ROW;
-089}
-090return ReturnCode.INCLUDE;
-091  }
-092
-093  @Override
-094  public boolean filterRowKey(Cell 
firstRowCell) {
-095if (compareRow(getCompareOperator(), 
this.comparator, firstRowCell)) {
-096  this.filterOutRow = true;
-097}
-098return this.filterOutRow;
-099  }
-100
-101  @Override
-102  public boolean filterRow() {
-103return this.filterOutRow;
-104  }
-105
-106  public static Filter 
createFilterFromArguments(ArrayListbyte [] filterArguments) {
-107@SuppressWarnings("rawtypes") // for 
arguments
-108ArrayList arguments = 
CompareFilter.extractArguments(filterArguments);
-109CompareOperator compareOp = 
(CompareOperator)arguments.get(0);
-110ByteArrayComparable comparator = 
(ByteArrayComparable)arguments.get(1);
-111return new RowFilter(compareOp, 
comparator);
-112  }
-113
-114 /**
-115  * @return The filter serialized using 
pb
-116  */
-117  @Override
-118  public byte [] toByteArray() {
-119FilterProtos.RowFilter.Builder 
builder =
-120  
FilterProtos.RowFilter.newBuilder();
-121
builder.setCompareFilter(super.convert());
-122return 
builder.build().toByteArray();
-123  }
-124
-125  /**
-126   * @param pbBytes A pb serialized 
{@link RowFilter} instance
-127   * @return An instance of {@link 
RowFilter} made from codebytes/code
-128   * @throws DeserializationException
-129   * @see #toByteArray
-130   */
-131  public static RowFilter parseFrom(final 
byte [] pbBytes)
-132  throws DeserializationException {
-133FilterProtos.RowFilter proto;
-134try {
-135  proto = 
FilterProtos.RowFilter.parseFrom(pbBytes);
-136} catch 
(InvalidProtocolBufferException e) {
-137  throw new 
DeserializationException(e);
-138}
-139final CompareOperator valueCompareOp 
=
-140  
CompareOperator.valueOf(proto.getCompareFilter().getCompareOp().name());
-141ByteArrayComparable valueComparator = 
null;
-142try {
-143  if 
(proto.getCompareFilter().hasComparator()) {
-144valueComparator = 

[27/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.

2018-08-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
index db8431b..a8cb7c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
@@ -885,7766 +885,7797 @@
 877   * @return What the next sequence 
(edit) id should be.
 878   * @throws IOException e
 879   */
-880  private long initialize(final 
CancelableProgressable reporter) throws IOException {
-881
-882//Refuse to open the region if there 
is no column family in the table
-883if 
(htableDescriptor.getColumnFamilyCount() == 0) {
-884  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
-885  " should have at least one 
column family.");
-886}
-887
-888MonitoredTask status = 
TaskMonitor.get().createStatus("Initializing region " + this);
-889long nextSeqId = -1;
-890try {
-891  nextSeqId = 
initializeRegionInternals(reporter, status);
-892  return nextSeqId;
-893} finally {
-894  // nextSeqid will be -1 if the 
initialization fails.
-895  // At least it will be 0 
otherwise.
-896  if (nextSeqId == -1) {
-897status.abort("Exception during 
region " + getRegionInfo().getRegionNameAsString() +
-898  " initialization.");
-899  }
-900}
-901  }
-902
-903  private long 
initializeRegionInternals(final CancelableProgressable reporter,
-904  final MonitoredTask status) throws 
IOException {
-905if (coprocessorHost != null) {
-906  status.setStatus("Running 
coprocessor pre-open hook");
-907  coprocessorHost.preOpen();
-908}
-909
-910// Write HRI to a file in case we 
need to recover hbase:meta
-911// Only the primary replica should 
write .regioninfo
-912if 
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
-913  status.setStatus("Writing region 
info on filesystem");
-914  fs.checkRegionInfoOnFilesystem();
-915}
-916
-917// Initialize all the HStores
-918status.setStatus("Initializing all 
the Stores");
-919long maxSeqId = 
initializeStores(reporter, status);
-920this.mvcc.advanceTo(maxSeqId);
-921if 
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-922  CollectionHStore stores = 
this.stores.values();
-923  try {
-924// update the stores that we are 
replaying
-925LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
-926
stores.forEach(HStore::startReplayingFromWAL);
-927// Recover any edits if 
available.
-928maxSeqId = Math.max(maxSeqId,
-929  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-930// Make sure mvcc is up to max.
-931this.mvcc.advanceTo(maxSeqId);
-932  } finally {
-933LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
-934// update the stores that we are 
done replaying
-935
stores.forEach(HStore::stopReplayingFromWAL);
-936  }
-937}
-938this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-939
-940
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-941this.writestate.flushRequested = 
false;
-942this.writestate.compacting.set(0);
-943
-944if (this.writestate.writesEnabled) 
{
-945  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
-946  // Remove temporary data left over 
from old regions
-947  status.setStatus("Cleaning up 
temporary data from old regions");
-948  fs.cleanupTempDir();
-949}
-950
-951if (this.writestate.writesEnabled) 
{
-952  status.setStatus("Cleaning up 
detritus from prior splits");
-953  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-954  // these directories here on open.  
We may be opening a region that was
-955  // being split but we crashed in 
the middle of it all.
-956  LOG.debug("Cleaning up detritus for 
" + this.getRegionInfo().getEncodedName());
-957  fs.cleanupAnySplitDetritus();
-958  fs.cleanupMergesDir();
-959}
+880  @VisibleForTesting
+881  long initialize(final 
CancelableProgressable reporter) throws IOException {
+882
+883//Refuse to open the region if there 
is no column family in the table
+884if 
(htableDescriptor.getColumnFamilyCount() == 0) {
+885  throw new 
DoNotRetryIOException("Table " + 
htableDescriptor.getTableName().getNameAsString()+
+886  " should have at least one 
column family.");
+887}

[27/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.

2018-08-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
index 8ffeb95..b1fbd27 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":6,"i5":10,"i6":10,"i7":10,"i8":10,"i9":6,"i10":10,"i11":10,"i12":10,"i13":10,"i14":6,"i15":10,"i16":10,"i17":10,"i18":6,"i19":6,"i20":10,"i21":10,"i22":10,"i23":6,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":6,"i32":10,"i33":10,"i34":6,"i35":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":42,"i1":42,"i2":38,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":38,"i10":38,"i11":38,"i12":42,"i13":42,"i14":42,"i15":38,"i16":42,"i17":42,"i18":42,"i19":42,"i20":38,"i21":42,"i22":42,"i23":38};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
 
 
 PrevClass
-NextClass
+NextClass
 
 
 Frames
@@ -121,58 +121,20 @@ var activeTableTab = "activeTableTab";
 AssignProcedure, UnassignProcedure
 
 
+Deprecated.
+Do not use any 
more.
+
 
-@InterfaceAudience.Private
-public abstract class RegionTransitionProcedure
+https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true;
 title="class or interface in java.lang">@Deprecated
+ @InterfaceAudience.Private
+public abstract class RegionTransitionProcedure
 extends ProcedureMasterProcedureEnv
 implements TableProcedureInterface, RemoteProcedureDispatcher.RemoteProcedureMasterProcedureEnv,ServerName
-Base class for the Assign and Unassign Procedure.
-
- Locking:
- Takes exclusive lock on the region being assigned/unassigned. Thus, there can 
only be one
- RegionTransitionProcedure per region running at a time (see 
MasterProcedureScheduler).
-
- This procedure is asynchronous and responds to external events.
- The AssignmentManager will notify this procedure when the RS completes
- the operation and reports the transitioned state
- (see the Assign and Unassign class for more detail).
-
- Procedures move from the REGION_TRANSITION_QUEUE state when they are
- first submitted, to the REGION_TRANSITION_DISPATCH state when the request
- to remote server is sent and the Procedure is suspended waiting on external
- event to be woken again. Once the external event is triggered, Procedure
- moves to the REGION_TRANSITION_FINISH state.
-
- NOTE: AssignProcedure and 
UnassignProcedure 
should not be thought of
- as being asymmetric, at least currently.
- 
- AssignProcedure 
moves through all the above described states and implements methods
- associated with each while UnassignProcedure 
starts at state
- REGION_TRANSITION_DISPATCH and state REGION_TRANSITION_QUEUE is not 
supported.
-
- When any step in AssignProcedure 
fails, failure handler
- AssignProcedure#handleFailure(MasterProcedureEnv, RegionStateNode) 
re-attempts the
- assignment by setting the procedure state to REGION_TRANSITION_QUEUE and 
forces
- assignment to a different target server by setting AssignProcedure.forceNewPlan.
 When
- the number of attempts reaches threshold configuration 
'hbase.assignment.maximum.attempts',
- the procedure is aborted. For UnassignProcedure, 
similar re-attempts are
- intentionally not implemented. It is a 'one shot' procedure. See its class 
doc for how it
- handles failure.
- 
- If we find a region in an 'unexpected' state, we'll complain and retry 
with backoff forever.
- The 'unexpected' state needs to be fixed either by another running Procedure 
or by operator
- intervention (Regions in 'unexpected' state indicates bug or unexpected 
transition type).
- For this to work, subclasses need to persist the 'attempt' counter kept in 
this class when
- they do serializeStateData and restore it inside their deserializeStateData, 
just as they do
- for regionInfo.
- 
- 
- 
-
- TODO: Considering it is a priority doing all we can to get make a region 
available as soon as
- possible, re-attempting with any target makes sense if specified target fails 
in case of
- AssignProcedure. 
For UnassignProcedure, 
our concern is preventing data loss
- on failed unassign. See class doc for explanation.
+Leave here only for checking if we can successfully start 
the master.
+
+See Also:

[27/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.

2018-08-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
index 39170f0..7859ebc 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html
@@ -230,564 +230,567 @@
 222  }
 223}
 224  } catch (InterruptedException e) 
{
-225e.printStackTrace();
-226  }
-227}
-228
-229@Override
-230public void setup(Context context) 
throws IOException {
-231  cfRenameMap = 
createCfRenameMap(context.getConfiguration());
-232  filter = 
instantiateFilter(context.getConfiguration());
-233  int reduceNum = 
context.getNumReduceTasks();
-234  Configuration conf = 
context.getConfiguration();
-235  TableName tableName = 
TableName.valueOf(context.getConfiguration().get(TABLE_NAME));
-236  try (Connection conn = 
ConnectionFactory.createConnection(conf);
-237  RegionLocator regionLocator = 
conn.getRegionLocator(tableName)) {
-238byte[][] startKeys = 
regionLocator.getStartKeys();
-239if (startKeys.length != 
reduceNum) {
-240  throw new IOException("Region 
split after job initialization");
-241}
-242CellWritableComparable[] 
startKeyWraps =
-243new 
CellWritableComparable[startKeys.length - 1];
-244for (int i = 1; i  
startKeys.length; ++i) {
-245  startKeyWraps[i - 1] =
-246  new 
CellWritableComparable(KeyValueUtil.createFirstOnRow(startKeys[i]));
-247}
-248
CellWritableComparablePartitioner.START_KEYS = startKeyWraps;
-249  }
-250}
-251  }
-252
-253  /**
-254   * A mapper that just writes out 
KeyValues.
-255   */
-256  
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="EQ_COMPARETO_USE_OBJECT_EQUALS",
-257  justification="Writables are going 
away and this has been this way forever")
-258  public static class CellImporter 
extends TableMapperImmutableBytesWritable, Cell {
-259private Mapbyte[], byte[] 
cfRenameMap;
-260private Filter filter;
-261private static final Logger LOG = 
LoggerFactory.getLogger(CellImporter.class);
-262
-263/**
-264 * @param row  The current table row 
key.
-265 * @param value  The columns.
-266 * @param context  The current 
context.
-267 * @throws IOException When something 
is broken with the data.
-268 */
-269@Override
-270public void 
map(ImmutableBytesWritable row, Result value,
-271  Context context)
-272throws IOException {
-273  try {
-274if (LOG.isTraceEnabled()) {
-275  LOG.trace("Considering the 
row."
-276  + Bytes.toString(row.get(), 
row.getOffset(), row.getLength()));
-277}
-278if (filter == null
-279|| 
!filter.filterRowKey(PrivateCellUtil.createFirstOnRow(row.get(), 
row.getOffset(),
-280(short) 
row.getLength( {
-281  for (Cell kv : 
value.rawCells()) {
-282kv = filterKv(filter, kv);
-283// skip if we filtered it 
out
-284if (kv == null) continue;
-285context.write(row, new 
MapReduceExtendedCell(convertKv(kv, cfRenameMap)));
-286  }
-287}
-288  } catch (InterruptedException e) 
{
-289e.printStackTrace();
-290  }
-291}
-292
-293@Override
-294public void setup(Context context) 
{
-295  cfRenameMap = 
createCfRenameMap(context.getConfiguration());
-296  filter = 
instantiateFilter(context.getConfiguration());
-297}
-298  }
-299
-300  /**
-301   * Write table content out to files in 
hdfs.
-302   */
-303  public static class Importer extends 
TableMapperImmutableBytesWritable, Mutation {
-304private Mapbyte[], byte[] 
cfRenameMap;
-305private ListUUID 
clusterIds;
-306private Filter filter;
-307private Durability durability;
-308
-309/**
-310 * @param row  The current table row 
key.
-311 * @param value  The columns.
-312 * @param context  The current 
context.
-313 * @throws IOException When something 
is broken with the data.
-314 */
-315@Override
-316public void 
map(ImmutableBytesWritable row, Result value,
-317  Context context)
-318throws IOException {
-319  try {
-320writeResult(row, value, 
context);
-321  } catch (InterruptedException e) 
{
-322e.printStackTrace();
-323  }
-324}
-325
-326private void 
writeResult(ImmutableBytesWritable key, Result result, Context context)
-327throws IOException, 

[27/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.

2018-08-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html
index b7b4236..3d1edb3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html
@@ -259,1863 +259,1867 @@
 251   * + Metadata!  + = See note on 
BLOCK_METADATA_SPACE above.
 252   * ++
 253   * /code
-254   * @see #serialize(ByteBuffer)
+254   * @see #serialize(ByteBuffer, 
boolean)
 255   */
-256  static final 
CacheableDeserializerCacheable BLOCK_DESERIALIZER =
-257  new 
CacheableDeserializerCacheable() {
-258@Override
-259public HFileBlock 
deserialize(ByteBuff buf, boolean reuse, MemoryType memType)
-260throws IOException {
-261  // The buf has the file block 
followed by block metadata.
-262  // Set limit to just before the 
BLOCK_METADATA_SPACE then rewind.
-263  buf.limit(buf.limit() - 
BLOCK_METADATA_SPACE).rewind();
-264  // Get a new buffer to pass the 
HFileBlock for it to 'own'.
-265  ByteBuff newByteBuff;
-266  if (reuse) {
-267newByteBuff = buf.slice();
-268  } else {
-269int len = buf.limit();
-270newByteBuff = new 
SingleByteBuff(ByteBuffer.allocate(len));
-271newByteBuff.put(0, buf, 
buf.position(), len);
-272  }
-273  // Read out the 
BLOCK_METADATA_SPACE content and shove into our HFileBlock.
-274  buf.position(buf.limit());
-275  buf.limit(buf.limit() + 
HFileBlock.BLOCK_METADATA_SPACE);
-276  boolean usesChecksum = buf.get() == 
(byte) 1;
-277  long offset = buf.getLong();
-278  int nextBlockOnDiskSize = 
buf.getInt();
-279  HFileBlock hFileBlock =
-280  new HFileBlock(newByteBuff, 
usesChecksum, memType, offset, nextBlockOnDiskSize, null);
-281  return hFileBlock;
-282}
-283
-284@Override
-285public int 
getDeserialiserIdentifier() {
-286  return DESERIALIZER_IDENTIFIER;
-287}
-288
-289@Override
-290public HFileBlock 
deserialize(ByteBuff b) throws IOException {
-291  // Used only in tests
-292  return deserialize(b, false, 
MemoryType.EXCLUSIVE);
-293}
-294  };
-295
-296  private static final int 
DESERIALIZER_IDENTIFIER;
-297  static {
-298DESERIALIZER_IDENTIFIER =
-299
CacheableDeserializerIdManager.registerDeserializer(BLOCK_DESERIALIZER);
-300  }
-301
-302  /**
-303   * Copy constructor. Creates a shallow 
copy of {@code that}'s buffer.
-304   */
-305  private HFileBlock(HFileBlock that) {
-306this(that, false);
-307  }
-308
-309  /**
-310   * Copy constructor. Creates a 
shallow/deep copy of {@code that}'s buffer as per the boolean
-311   * param.
-312   */
-313  private HFileBlock(HFileBlock that, 
boolean bufCopy) {
-314init(that.blockType, 
that.onDiskSizeWithoutHeader,
-315
that.uncompressedSizeWithoutHeader, that.prevBlockOffset,
-316that.offset, 
that.onDiskDataSizeWithHeader, that.nextBlockOnDiskSize, that.fileContext);
-317if (bufCopy) {
-318  this.buf = new 
SingleByteBuff(ByteBuffer.wrap(that.buf.toBytes(0, that.buf.limit(;
-319} else {
-320  this.buf = that.buf.duplicate();
-321}
-322  }
-323
-324  /**
-325   * Creates a new {@link HFile} block 
from the given fields. This constructor
-326   * is used only while writing blocks 
and caching,
-327   * and is sitting in a byte buffer and 
we want to stuff the block into cache.
-328   *
-329   * pTODO: The caller presumes 
no checksumming
-330   * required of this block instance 
since going into cache; checksum already verified on
-331   * underlying block data pulled in from 
filesystem. Is that correct? What if cache is SSD?
+256  public static final 
CacheableDeserializerCacheable BLOCK_DESERIALIZER = new 
BlockDeserializer();
+257
+258  public static final class 
BlockDeserializer implements CacheableDeserializerCacheable {
+259private BlockDeserializer() {
+260}
+261
+262@Override
+263public HFileBlock 
deserialize(ByteBuff buf, boolean reuse, MemoryType memType)
+264throws IOException {
+265  // The buf has the file block 
followed by block metadata.
+266  // Set limit to just before the 
BLOCK_METADATA_SPACE then rewind.
+267  buf.limit(buf.limit() - 
BLOCK_METADATA_SPACE).rewind();
+268  // Get a new buffer to pass the 
HFileBlock for it to 'own'.
+269  ByteBuff newByteBuff;
+270  if (reuse) {
+271newByteBuff = buf.slice();
+272  } else {
+273int len = buf.limit();
+274newByteBuff = new 
SingleByteBuff(ByteBuffer.allocate(len));
+275newByteBuff.put(0, buf, 
buf.position(), len);
+276  }
+277  

[27/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.

2018-07-25 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
index 9d1748f..ac41873 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html
@@ -411,7 +411,7 @@ extends AbstractStateMachineTableProcedure
-acquireLock,
 checkOnline,
 checkTableModifiable,
 getRegionDir,
 getUser, preflightChecks,
 releaseLock,
 releaseSyncLatch,
 setUser
+acquireLock,
 checkOnline,
 checkTableModifiable,
 getRegionDir,
 getUser, preflightChecks,
 releaseLock,
 releaseSyncLatch,
 setUser,
 waitInitialized<
 /li>
 
 
 
@@ -425,7 +425,7 @@ extends Procedure
-addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 doExecute,
 d
 oReleaseLock, doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics, getProcId,
 getProcIdHashCode,
 getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState, getSubmittedTime,
 getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout, haveSameParent,
 holdLock,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 <
 a 
href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#isWaiting--">isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey, setOwner,
 setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB, updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
+addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doExecute,
 doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics,
 getProcId,
 getProcIdHashCode, getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime,
 getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 holdLock, 
incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner,
 setOwner, setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout, setTimeoutFailure,
 shouldWaitClientAck,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB,
 updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html
index ad95213..941b155 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html
@@ -411,7 +411,7 @@ implements Procedure
-addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 doExecute,
 d
 oReleaseLock, doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcId, getProcIdHashCode,
 getProcName,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime,
  getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 holdLock,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner, setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB,
 updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
+addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doExecute,
 doRollback,
 

[27/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.

2018-07-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
index 233dba3..91b9055 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html
@@ -540,1205 +540,1204 @@
 532  
sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) - {
 533DequeBalancerRegionLoad 
rLoads = oldLoads.get(Bytes.toString(regionName));
 534if (rLoads == null) {
-535  // There was nothing there
-536  rLoads = new 
ArrayDeque();
-537} else if (rLoads.size() = 
numRegionLoadsToRemember) {
-538  rLoads.remove();
-539}
-540rLoads.add(new 
BalancerRegionLoad(rm));
-541
loads.put(Bytes.toString(regionName), rLoads);
-542  });
-543});
-544
-545for(CostFromRegionLoadFunction cost : 
regionLoadFunctions) {
-546  cost.setLoads(loads);
-547}
-548  }
-549
-550  protected void initCosts(Cluster 
cluster) {
-551for (CostFunction c:costFunctions) 
{
-552  c.init(cluster);
-553}
-554  }
-555
-556  protected void 
updateCostsWithAction(Cluster cluster, Action action) {
-557for (CostFunction c : costFunctions) 
{
-558  c.postAction(action);
-559}
-560  }
-561
-562  /**
-563   * Get the names of the cost 
functions
-564   */
-565  public String[] getCostFunctionNames() 
{
-566if (costFunctions == null) return 
null;
-567String[] ret = new 
String[costFunctions.length];
-568for (int i = 0; i  
costFunctions.length; i++) {
-569  CostFunction c = 
costFunctions[i];
-570  ret[i] = 
c.getClass().getSimpleName();
-571}
-572
-573return ret;
-574  }
-575
-576  /**
-577   * This is the main cost function.  It 
will compute a cost associated with a proposed cluster
-578   * state.  All different costs will be 
combined with their multipliers to produce a double cost.
-579   *
-580   * @param cluster The state of the 
cluster
-581   * @param previousCost the previous 
cost. This is used as an early out.
-582   * @return a double of a cost 
associated with the proposed cluster state.  This cost is an
-583   * aggregate of all individual 
cost functions.
-584   */
-585  protected double computeCost(Cluster 
cluster, double previousCost) {
-586double total = 0;
-587
-588for (int i = 0; i  
costFunctions.length; i++) {
-589  CostFunction c = 
costFunctions[i];
-590  this.tempFunctionCosts[i] = 0.0;
-591
-592  if (c.getMultiplier() = 0) {
-593continue;
-594  }
-595
-596  Float multiplier = 
c.getMultiplier();
-597  Double cost = c.cost();
-598
-599  this.tempFunctionCosts[i] = 
multiplier*cost;
-600  total += 
this.tempFunctionCosts[i];
-601
-602  if (total  previousCost) {
-603break;
-604  }
-605}
-606
-607return total;
-608  }
-609
-610  /** Generates a candidate action to be 
applied to the cluster for cost function search */
-611  abstract static class 
CandidateGenerator {
-612abstract Cluster.Action 
generate(Cluster cluster);
-613
-614/**
-615 * From a list of regions pick a 
random one. Null can be returned which
-616 * {@link 
StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region 
move
-617 * rather than swap.
-618 *
-619 * @param clusterThe state of 
the cluster
-620 * @param server index of the 
server
-621 * @param chanceOfNoSwap Chance that 
this will decide to try a move rather
-622 *   than a 
swap.
-623 * @return a random {@link 
RegionInfo} or null if an asymmetrical move is
-624 * suggested.
-625 */
-626protected int 
pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) {
-627  // Check to see if this is just a 
move.
-628  if 
(cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat()  
chanceOfNoSwap) {
-629// signal a move only.
-630return -1;
-631  }
-632  int rand = 
RANDOM.nextInt(cluster.regionsPerServer[server].length);
-633  return 
cluster.regionsPerServer[server][rand];
-634
-635}
-636protected int 
pickRandomServer(Cluster cluster) {
-637  if (cluster.numServers  1) {
-638return -1;
-639  }
-640
-641  return 
RANDOM.nextInt(cluster.numServers);
-642}
-643
-644protected int pickRandomRack(Cluster 
cluster) {
-645  if 

[27/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.

2018-07-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html 
b/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
index f1708c1..8a6ff31 100644
--- a/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
+++ b/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html
@@ -1,6 +1,6 @@
 http://www.w3.org/TR/html4/loose.dtd;>
 
-
+
 
 
 
@@ -19,45 +19,45 @@
 }
 //-->
 var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":9,"i15":10,"i16":9,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10};
-var tabs = 
{65535:["t0","所有方法"],1:["t1","静态方法"],2:["t2","实例方法"],8:["t4","å
…·ä½“方法"]};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
 var activeTableTab = "activeTableTab";
 
 
-您的浏览器已禁用 JavaScript。
+JavaScript is disabled on your browser.
 
 
 
 
 
-跳过导航链接
+Skip navigation links
 
 
 
-
-概览
-程序包
-ç±»
-使用
-树
-已过时
-索引
-帮助
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
 
 
 
 
-上一个类
-下一个类
+PrevClass
+NextClass
 
 
-框架
-无框架
+Frames
+NoFrames
 
 
-所有类
+AllClasses
 
 
 

[27/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.

2018-07-19 Thread zhangduo
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html
--
diff --git 
a/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html 
b/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html
index c0db59c..74ecf35 100644
--- a/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html
+++ b/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html
@@ -1,6 +1,6 @@
 http://www.w3.org/TR/html4/loose.dtd;>
 
-
+
 
 
 
@@ -20,38 +20,38 @@
 //-->
 
 
-JavaScript is disabled on your browser.
+您的浏览器已禁用 JavaScript。
 
 
 
 
 
-Skip navigation links
+跳过导航链接
 
 
 
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
+
+概览
+程序包
+ç±»
+使用
+树
+已过时
+索引
+帮助
 
 
 
 
-PrevClass
-NextClass
+上一个类
+下一个类
 
 
-Frames
-NoFrames
+框架
+无
框架
 
 
-AllClasses
+所有类
 
 
 
 
org.apache.hadoop.hbase
-

Class InvalidFamilyOperationException

+

ç±» InvalidFamilyOperationException


[27/51] [partial] hbase-site git commit: Published site at 0f23784182ab88649de340d75804e0ff20dcd0fc.

2018-07-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html
index e23e8f7..cf50fff 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10};
+var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -186,37 +186,29 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
longoffHeapSizeDelta)
 
 
-void
-blockUpdates()
-
-
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ThreadPoolExecutor.html?is-external=true;
 title="class or interface in 
java.util.concurrent">ThreadPoolExecutor
 getInMemoryCompactionPool()
 
-
+
 long
 getMemStoreFlushSize()
 
-
+
 (package private) long
 getMemStoreSize()
 
-
+
 int
 getNumStores()
 
-
+
 RegionInfo
 getRegionInfo()
 
-
+
 WAL
 getWAL()
 
-
-void
-unblockUpdates()
-
 
 
 
@@ -295,31 +287,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 Method Detail
-
-
-
-
-
-blockUpdates
-publicvoidblockUpdates()
-
-
-
-
-
-
-
-unblockUpdates
-publicvoidunblockUpdates()
-
-
 
 
 
 
 
 addMemStoreSize
-publicvoidaddMemStoreSize(longdataSizeDelta,
+publicvoidaddMemStoreSize(longdataSizeDelta,
 longheapSizeDelta,
 longoffHeapSizeDelta)
 
@@ -330,7 +304,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getRegionInfo
-publicRegionInfogetRegionInfo()
+publicRegionInfogetRegionInfo()
 
 
 
@@ -339,7 +313,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getWAL
-publicWALgetWAL()
+publicWALgetWAL()
 
 
 
@@ -348,7 +322,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getInMemoryCompactionPool
-publichttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ThreadPoolExecutor.html?is-external=true;
 title="class or interface in 
java.util.concurrent">ThreadPoolExecutorgetInMemoryCompactionPool()
+publichttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ThreadPoolExecutor.html?is-external=true;
 title="class or interface in 
java.util.concurrent">ThreadPoolExecutorgetInMemoryCompactionPool()
 
 
 
@@ -357,7 +331,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getMemStoreFlushSize
-publiclonggetMemStoreFlushSize()
+publiclonggetMemStoreFlushSize()
 
 
 
@@ -366,7 +340,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getNumStores
-publicintgetNumStores()
+publicintgetNumStores()
 
 
 
@@ -375,7 +349,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 getMemStoreSize
-longgetMemStoreSize()
+longgetMemStoreSize()
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html
index 84a1668..5d9b1ff 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = {"i0":6,"i1":6,"i2":6};
+var methods = {"i0":6,"i1":6,"i2":6,"i3":6};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public interface ReplicationSourceService
+public interface ReplicationSourceService
 extends ReplicationService
 A source for a replication stream has to expose this 
service. This service allows an application
  to hook into the regionserver and watch for new transactions.
@@ -139,12 +139,18 @@ extends 
+ReplicationSourceManager
+getReplicationManager()
+Returns the replication manager
+
+
+
 ReplicationPeers
 getReplicationPeers()
 Return the replication peers.
 
 
-
+
 SyncReplicationPeerInfoProvider
 getSyncReplicationPeerInfoProvider()
 Returns an info provider for sync replication peer.
@@ -178,7 +184,7 @@ extends 
 
 

[27/51] [partial] hbase-site git commit: Published site at 85b41f36e01214b6485c9352875c84ebf877dab3.

2018-06-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html 
b/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html
index 2b98dff..a644a4a 100644
--- a/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html
+++ b/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9};
+var methods = 
{"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class AccessControlUtil
+public class AccessControlUtil
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 
 Since:
@@ -229,13 +229,42 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListUserPermission
+getUserPermissions(com.google.protobuf.RpcControllercontroller,
+  
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol,
+  byte[]namespace,
+  https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringuserName)
+A utility used to get permissions for selected namespace 
based on the specified user name.
+
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListUserPermission
+getUserPermissions(com.google.protobuf.RpcControllercontroller,
+  
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol,
+  https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringuserName)
+A utility used to get user's global permissions based on 
the specified user name.
+
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListUserPermission
 getUserPermissions(com.google.protobuf.RpcControllercontroller,
   
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol,
   TableNamet)
 A utility used to get user table permissions.
 
 
-
+
+static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListUserPermission
+getUserPermissions(com.google.protobuf.RpcControllercontroller,
+  
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol,
+  TableNamet,
+  byte[]columnFamily,
+  byte[]columnQualifier,
+  https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringuserName)
+A utility used to get user table permissions based on the 
column family, column qualifier and
+ user name.
+
+
+
 static void
 grant(com.google.protobuf.RpcControllercontroller,
  
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol,
@@ -245,7 +274,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 A utility used to grant a user global permissions.
 
 
-
+
 static void
 grant(com.google.protobuf.RpcControllercontroller,
  
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol,
@@ -256,7 +285,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 A utility used to grant a user namespace permissions.
 
 
-
+
 static void
 grant(com.google.protobuf.RpcControllercontroller,
  
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol,
@@ -269,7 +298,20 @@ extends 

[27/51] [partial] hbase-site git commit: Published site at 6198e1fc7dfa85c3bc6b2855f9a5fb5f4b2354ff.

2018-06-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
index 1410333..b56a22f 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html
@@ -124,7 +124,7 @@ var activeTableTab = "activeTableTab";
 
 @InterfaceAudience.Private
  @InterfaceStability.Evolving
-private static class HBaseAdmin.ReplicationFuture
+private static class HBaseAdmin.ReplicationFuture
 extends HBaseAdmin.ProcedureFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
 
 
@@ -241,7 +241,7 @@ extends 
 
 peerId
-private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String peerId
+private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String peerId
 
 
 
@@ -250,7 +250,7 @@ extends 
 
 getOperation
-private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true;
 title="class or interface in java.util.function">Supplierhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String getOperation
+private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true;
 title="class or interface in java.util.function">Supplierhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String getOperation
 
 
 
@@ -267,7 +267,7 @@ extends 
 
 ReplicationFuture
-publicReplicationFuture(HBaseAdminadmin,
+publicReplicationFuture(HBaseAdminadmin,
  https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringpeerId,
  https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">LongprocId,
  https://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true;
 title="class or interface in java.util.function">Supplierhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetOperation)
@@ -287,7 +287,7 @@ extends 
 
 toString
-publichttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtoString()
+publichttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtoString()
 
 Overrides:
 https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
 title="class or interface in java.lang">toStringin 
classhttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
index 9732695..356 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
@@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HBaseAdmin.RestoreSnapshotFuture
+private static class HBaseAdmin.RestoreSnapshotFuture
 extends HBaseAdmin.TableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
 
 
@@ -240,7 +240,7 @@ extends 
 
 RestoreSnapshotFuture
-publicRestoreSnapshotFuture(HBaseAdminadmin,
+publicRestoreSnapshotFuture(HBaseAdminadmin,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot,
  TableNametableName,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponseresponse)
@@ -252,7 +252,7 @@ extends 
 
 RestoreSnapshotFuture
-publicRestoreSnapshotFuture(HBaseAdminadmin,
+publicRestoreSnapshotFuture(HBaseAdminadmin,
  TableNametableName,
  

[27/51] [partial] hbase-site git commit: Published site at 14087cc919da9f2e0b1a68f701f6365ad9d1d71f.

2018-06-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
index b0c0fe7..08be9d0 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html
@@ -86,6 +86,10 @@
 org.apache.hadoop.hbase.master.assignment
 
 
+
+org.apache.hadoop.hbase.master.procedure
+
+
 
 
 
@@ -103,12 +107,27 @@
 
 
 
-MoveRegionProcedure
+private MoveRegionProcedure
 AssignmentManager.createMoveRegionProcedure(RegionPlanplan)
 
-
-MoveRegionProcedure[]
-AssignmentManager.createReopenProcedures(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionRegionInforegionInfo)
+
+
+
+
+
+
+Uses of MoveRegionProcedure in org.apache.hadoop.hbase.master.procedure
+
+Methods in org.apache.hadoop.hbase.master.procedure
 that return MoveRegionProcedure
+
+Modifier and Type
+Method and Description
+
+
+
+private MoveRegionProcedure
+ReopenTableRegionsProcedure.createReopenProcedure(MasterProcedureEnvenv,
+ HRegionLocationloc)
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
index f9b3413..a8b7714 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html
@@ -245,40 +245,44 @@
 RegionStates.RegionStateNode.compareTo(RegionStates.RegionStateNodeother)
 
 
+private HRegionLocation
+RegionStates.createRegionForReopen(RegionStates.RegionStateNodenode)
+
+
 protected void
 UnassignProcedure.finishTransition(MasterProcedureEnvenv,
 RegionStates.RegionStateNoderegionNode)
 
-
+
 protected void
 AssignProcedure.finishTransition(MasterProcedureEnvenv,
 RegionStates.RegionStateNoderegionNode)
 
-
+
 protected abstract void
 RegionTransitionProcedure.finishTransition(MasterProcedureEnvenv,
 RegionStates.RegionStateNoderegionNode)
 
-
+
 private void
 AssignProcedure.handleFailure(MasterProcedureEnvenv,
  RegionStates.RegionStateNoderegionNode)
 Called when dispatch or subsequent OPEN request fail.
 
 
-
+
 (package private) boolean
 RegionStates.include(RegionStates.RegionStateNodenode,
booleanoffline)
 Utility.
 
 
-
+
 private boolean
 AssignProcedure.incrementAndCheckMaxAttempts(MasterProcedureEnvenv,
 RegionStates.RegionStateNoderegionNode)
 
-
+
 private boolean
 UnassignProcedure.isSafeToProceed(MasterProcedureEnvenv,
RegionStates.RegionStateNoderegionNode,
@@ -287,143 +291,143 @@
  unassign; e.g.
 
 
-
+
 protected boolean
 RegionTransitionProcedure.isServerOnline(MasterProcedureEnvenv,
   RegionStates.RegionStateNoderegionNode)
 
-
+
 void
 AssignmentManager.markRegionAsClosed(RegionStates.RegionStateNoderegionNode)
 
-
+
 void
 AssignmentManager.markRegionAsClosing(RegionStates.RegionStateNoderegionNode)
 
-
+
 void
 AssignmentManager.markRegionAsOpened(RegionStates.RegionStateNoderegionNode)
 
-
+
 void
 AssignmentManager.markRegionAsOpening(RegionStates.RegionStateNoderegionNode)
 
-
+
 protected void
 UnassignProcedure.proceed(MasterProcedureEnvenv,
RegionStates.RegionStateNoderegionNode)
 Set it up so when procedure is unsuspended, we'll move to 
the procedure finish.
 
 
-
+
 protected void
 AssignmentManager.queueAssign(RegionStates.RegionStateNoderegionNode)
 Add the assign operation to the assignment queue.
 
 
-
+
 protected boolean
 UnassignProcedure.remoteCallFailed(MasterProcedureEnvenv,
 RegionStates.RegionStateNoderegionNode,
 https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in 
java.io">IOExceptionexception)
 
-
+
 protected boolean
 AssignProcedure.remoteCallFailed(MasterProcedureEnvenv,
 RegionStates.RegionStateNoderegionNode,
 https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in 
java.io">IOExceptionexception)
 
-
+
 protected abstract boolean
 

[27/51] [partial] hbase-site git commit: Published site at 72784c2d836a4b977667449d3adec5e8d15453f5.

2018-06-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
index b6e7636..592c2cc 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html
@@ -356,3901 +356,3924 @@
 348  public FutureVoid 
modifyTableAsync(TableDescriptor td) throws IOException {
 349ModifyTableResponse response = 
executeCallable(
 350  new 
MasterCallableModifyTableResponse(getConnection(), 
getRpcControllerFactory()) {
-351@Override
-352protected ModifyTableResponse 
rpcCall() throws Exception {
-353  
setPriority(td.getTableName());
-354  ModifyTableRequest request = 
RequestConverter.buildModifyTableRequest(
-355td.getTableName(), td, 
ng.getNonceGroup(), ng.newNonce());
-356  return 
master.modifyTable(getRpcController(), request);
-357}
-358  });
-359return new ModifyTableFuture(this, 
td.getTableName(), response);
-360  }
-361
-362  @Override
-363  public ListTableDescriptor 
listTableDescriptorsByNamespace(byte[] name) throws IOException {
-364return executeCallable(new 
MasterCallableListTableDescriptor(getConnection(),
-365getRpcControllerFactory()) {
-366  @Override
-367  protected 
ListTableDescriptor rpcCall() throws Exception {
-368return 
master.listTableDescriptorsByNamespace(getRpcController(),
-369
ListTableDescriptorsByNamespaceRequest.newBuilder()
-370  
.setNamespaceName(Bytes.toString(name)).build())
-371.getTableSchemaList()
-372.stream()
-373
.map(ProtobufUtil::toTableDescriptor)
-374
.collect(Collectors.toList());
-375  }
-376});
-377  }
-378
-379  @Override
-380  public ListTableDescriptor 
listTableDescriptors(ListTableName tableNames) throws IOException {
-381return executeCallable(new 
MasterCallableListTableDescriptor(getConnection(),
-382getRpcControllerFactory()) {
-383  @Override
-384  protected 
ListTableDescriptor rpcCall() throws Exception {
-385GetTableDescriptorsRequest req 
=
-386
RequestConverter.buildGetTableDescriptorsRequest(tableNames);
-387  return 
ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(),
-388  req));
-389  }
-390});
-391  }
-392
-393  @Override
-394  public ListRegionInfo 
getRegions(final ServerName sn) throws IOException {
-395AdminService.BlockingInterface admin 
= this.connection.getAdmin(sn);
-396// TODO: There is no timeout on this 
controller. Set one!
-397HBaseRpcController controller = 
rpcControllerFactory.newController();
-398return 
ProtobufUtil.getOnlineRegions(controller, admin);
-399  }
-400
-401  @Override
-402  public ListRegionInfo 
getRegions(TableName tableName) throws IOException {
-403if 
(TableName.isMetaTableName(tableName)) {
-404  return 
Arrays.asList(RegionInfoBuilder.FIRST_META_REGIONINFO);
-405} else {
-406  return 
MetaTableAccessor.getTableRegions(connection, tableName, true);
-407}
-408  }
-409
-410  private static class 
AbortProcedureFuture extends ProcedureFutureBoolean {
-411private boolean isAbortInProgress;
-412
-413public AbortProcedureFuture(
-414final HBaseAdmin admin,
-415final Long procId,
-416final Boolean abortProcResponse) 
{
-417  super(admin, procId);
-418  this.isAbortInProgress = 
abortProcResponse;
-419}
-420
-421@Override
-422public Boolean get(long timeout, 
TimeUnit unit)
-423throws InterruptedException, 
ExecutionException, TimeoutException {
-424  if (!this.isAbortInProgress) {
-425return false;
-426  }
-427  super.get(timeout, unit);
-428  return true;
-429}
-430  }
-431
-432  /** @return Connection used by this 
object. */
-433  @Override
-434  public Connection getConnection() {
-435return connection;
-436  }
-437
-438  @Override
-439  public boolean tableExists(final 
TableName tableName) throws IOException {
-440return executeCallable(new 
RpcRetryingCallableBoolean() {
-441  @Override
-442  protected Boolean rpcCall(int 
callTimeout) throws Exception {
-443return 
MetaTableAccessor.tableExists(connection, tableName);
-444  }
-445});
-446  }
-447
-448  @Override
-449  public HTableDescriptor[] listTables() 
throws IOException {
-450return listTables((Pattern)null, 
false);
-451  }
-452
-453  @Override
-454  public HTableDescriptor[] 
listTables(Pattern pattern) throws IOException {

[27/51] [partial] hbase-site git commit: Published site at 9101fc246f86445006bfbcdfda5cc495016dc280.

2018-06-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html 
b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
index a9d814d..e65cf97 100644
--- a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
+++ b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -288,13 +288,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 private https://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true;
 title="class or interface in java.lang">ThreadGroup
 threadGroup
-Created in the start(int,
 boolean) method.
+Created in the #start(int, boolean) 
method.
 
 
 
 private TimeoutExecutorThread
 timeoutExecutor
-Created in the start(int,
 boolean) method.
+Created in the #start(int, boolean) 
method.
 
 
 
@@ -308,7 +308,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
 
 private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CopyOnWriteArrayList.html?is-external=true;
 title="class or interface in 
java.util.concurrent">CopyOnWriteArrayListProcedureExecutor.WorkerThread
 workerThreads
-Created in the start(int,
 boolean) method.
+Created in the #start(int, boolean) 
method.
 
 
 
@@ -490,107 +490,114 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
   https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
 title="class or interface in 
java.lang">InterruptedExceptione)
 
 
+void
+init(intnumThreads,
+booleanabortOnCorruption)
+Initialize the procedure executor, but do not start 
workers.
+
+
+
 private Procedure[]
 initializeChildren(RootProcedureStateprocStack,
   Procedureprocedure,
   Procedure[]subprocs)
 
-
+
 boolean
 isFinished(longprocId)
 Return true if the procedure is finished.
 
 
-
+
 boolean
 isProcedureOwner(longprocId,
 Useruser)
 Check if the user is this procedure's owner
 
 
-
+
 boolean
 isRunning()
 
-
+
 boolean
 isStarted(longprocId)
 Return true if the procedure is started.
 
 
-
+
 void
 join()
 
-
+
 private void
 load(booleanabortOnCorruption)
 
-
+
 private void
 loadProcedures(ProcedureStore.ProcedureIteratorprocIter,
   booleanabortOnCorruption)
 
-
+
 private long
 nextProcId()
 
-
+
 private Procedure
 prepareProcedure(Procedureproc)
 
-
+
 private void
 procedureFinished(Procedureproc)
 
-
+
 private long
 pushProcedure(Procedureproc)
 
-
+
 void
 refreshConfiguration(org.apache.hadoop.conf.Configurationconf)
 
-
+
 void
 registerListener(ProcedureExecutor.ProcedureExecutorListenerlistener)
 
-
+
 long
 registerNonce(NonceKeynonceKey)
 Register a nonce for a procedure that is going to be 
submitted.
 
 
-
+
 private void
 releaseLock(Procedureproc,
booleanforce)
 
-
+
 boolean
 removeChore(ProcedureInMemoryChorechore)
 Remove a chore procedure from the executor
 
 
-
+
 void
 removeResult(longprocId)
 Mark the specified completed procedure, as ready to 
remove.
 
 
-
+
 private void
 sendProcedureAddedNotification(longprocId)
 
-
+
 private void
 sendProcedureFinishedNotification(longprocId)
 
-
+
 private void
 sendProcedureLoadedNotification(longprocId)
 
-
+
 void
 setFailureResultForNonce(NonceKeynonceKey,
 https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 

[27/51] [partial] hbase-site git commit: Published site at 0b28155d274910b4e667b949d51f78809a1eff0b.

2018-06-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
index f236300..513d2ad 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
@@ -26,1048 +26,1115 @@
 018
 019package 
org.apache.hadoop.hbase.backup.impl;
 020
-021import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
-022import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
-023import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
-024import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
-025import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
-026import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
-027import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
-028import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
-029import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
-030import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
-031import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
-032import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
-033import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
-034import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
-035import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
-036import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
-037import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
-038import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
-039
-040import java.io.IOException;
-041import java.net.URI;
-042import java.util.List;
+021import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BACKUP_LIST_DESC;
+022import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
+023import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
+024import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
+025import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
+026import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP;
+027import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP_DESC;
+028import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_LIST;
+029import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
+030import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
+031import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
+032import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
+033import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
+034import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
+035import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
+036import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
+037import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
+038import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
+039import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
+040import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
+041import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
+042import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
 043
-044import 
org.apache.commons.lang3.StringUtils;
-045import 
org.apache.hadoop.conf.Configuration;
-046import 
org.apache.hadoop.conf.Configured;
-047import org.apache.hadoop.fs.FileSystem;
-048import org.apache.hadoop.fs.Path;
-049import 
org.apache.hadoop.hbase.HBaseConfiguration;
-050import 
org.apache.hadoop.hbase.TableName;
-051import 
org.apache.hadoop.hbase.backup.BackupAdmin;
-052import 
org.apache.hadoop.hbase.backup.BackupInfo;
-053import 

[27/51] [partial] hbase-site git commit: Published site at 7d3750bd9fc9747623549c242cc4171e224b3eaf.

2018-06-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
index 275ca84..b9a30c4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
@@ -1499,782 +1499,786 @@
 1491  final RpcController controller,
 1492  final IsInMaintenanceModeRequest 
request) throws ServiceException {
 1493IsInMaintenanceModeResponse.Builder 
response = IsInMaintenanceModeResponse.newBuilder();
-1494
response.setInMaintenanceMode(master.isInMaintenanceMode());
-1495return response.build();
-1496  }
-1497
-1498  @Override
-1499  public UnassignRegionResponse 
unassignRegion(RpcController controller,
-1500  UnassignRegionRequest req) throws 
ServiceException {
-1501try {
-1502  final byte [] regionName = 
req.getRegion().getValue().toByteArray();
-1503  RegionSpecifierType type = 
req.getRegion().getType();
-1504  final boolean force = 
req.getForce();
-1505  UnassignRegionResponse urr = 
UnassignRegionResponse.newBuilder().build();
-1506
-1507  master.checkInitialized();
-1508  if (type != 
RegionSpecifierType.REGION_NAME) {
-1509LOG.warn("unassignRegion 
specifier type: expected: " + RegionSpecifierType.REGION_NAME
-1510  + " actual: " + type);
-1511  }
-1512  PairRegionInfo, ServerName 
pair =
-1513
MetaTableAccessor.getRegion(master.getConnection(), regionName);
-1514  if 
(Bytes.equals(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(),regionName))
 {
-1515pair = new 
Pair(RegionInfoBuilder.FIRST_META_REGIONINFO,
-1516
master.getMetaTableLocator().getMetaRegionLocation(master.getZooKeeper()));
-1517  }
-1518  if (pair == null) {
-1519throw new 
UnknownRegionException(Bytes.toString(regionName));
-1520  }
-1521
-1522  RegionInfo hri = 
pair.getFirst();
-1523  if (master.cpHost != null) {
-1524master.cpHost.preUnassign(hri, 
force);
-1525  }
-1526  
LOG.debug(master.getClientIdAuditPrefix() + " unassign " + 
hri.getRegionNameAsString()
-1527  + " in current location if it 
is online and reassign.force=" + force);
-1528  
master.getAssignmentManager().unassign(hri);
-1529  if (master.cpHost != null) {
-1530master.cpHost.postUnassign(hri, 
force);
-1531  }
-1532
-1533  return urr;
-1534} catch (IOException ioe) {
-1535  throw new ServiceException(ioe);
-1536}
-1537  }
-1538
-1539  @Override
-1540  public 
ReportRegionStateTransitionResponse reportRegionStateTransition(RpcController 
c,
-1541  ReportRegionStateTransitionRequest 
req) throws ServiceException {
-1542try {
-1543  master.checkServiceStarted();
-1544  return 
master.getAssignmentManager().reportRegionStateTransition(req);
-1545} catch (IOException ioe) {
-1546  throw new ServiceException(ioe);
-1547}
-1548  }
-1549
-1550  @Override
-1551  public SetQuotaResponse 
setQuota(RpcController c, SetQuotaRequest req)
-1552  throws ServiceException {
-1553try {
-1554  master.checkInitialized();
-1555  return 
master.getMasterQuotaManager().setQuota(req);
-1556} catch (Exception e) {
-1557  throw new ServiceException(e);
-1558}
-1559  }
-1560
-1561  @Override
-1562  public 
MajorCompactionTimestampResponse getLastMajorCompactionTimestamp(RpcController 
controller,
-1563  MajorCompactionTimestampRequest 
request) throws ServiceException {
-1564
MajorCompactionTimestampResponse.Builder response =
-1565
MajorCompactionTimestampResponse.newBuilder();
-1566try {
-1567  master.checkInitialized();
-1568  
response.setCompactionTimestamp(master.getLastMajorCompactionTimestamp(ProtobufUtil
-1569  
.toTableName(request.getTableName(;
-1570} catch (IOException e) {
-1571  throw new ServiceException(e);
-1572}
-1573return response.build();
-1574  }
-1575
-1576  @Override
-1577  public 
MajorCompactionTimestampResponse getLastMajorCompactionTimestampForRegion(
-1578  RpcController controller, 
MajorCompactionTimestampForRegionRequest request)
-1579  throws ServiceException {
-1580
MajorCompactionTimestampResponse.Builder response =
-1581
MajorCompactionTimestampResponse.newBuilder();
-1582try {
-1583  master.checkInitialized();
-1584  
response.setCompactionTimestamp(master.getLastMajorCompactionTimestampForRegion(request
-1585  
.getRegion().getValue().toByteArray()));
-1586} catch (IOException e) {
-1587  throw new 

[27/51] [partial] hbase-site git commit: Published site at 997747076d8ec0b4346d7cb99c4b0667a7c14905.

2018-05-30 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
index 3da432b..d30fa8f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
@@ -928,7690 +928,7698 @@
 920  CollectionHStore stores = 
this.stores.values();
 921  try {
 922// update the stores that we are 
replaying
-923
stores.forEach(HStore::startReplayingFromWAL);
-924// Recover any edits if 
available.
-925maxSeqId = Math.max(maxSeqId,
-926  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
-927// Make sure mvcc is up to max.
-928this.mvcc.advanceTo(maxSeqId);
-929  } finally {
-930// update the stores that we are 
done replaying
-931
stores.forEach(HStore::stopReplayingFromWAL);
-932  }
-933}
-934this.lastReplayedOpenRegionSeqId = 
maxSeqId;
-935
-936
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-937this.writestate.flushRequested = 
false;
-938this.writestate.compacting.set(0);
-939
-940if (this.writestate.writesEnabled) 
{
-941  // Remove temporary data left over 
from old regions
-942  status.setStatus("Cleaning up 
temporary data from old regions");
-943  fs.cleanupTempDir();
-944}
-945
-946if (this.writestate.writesEnabled) 
{
-947  status.setStatus("Cleaning up 
detritus from prior splits");
-948  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
-949  // these directories here on open.  
We may be opening a region that was
-950  // being split but we crashed in 
the middle of it all.
-951  fs.cleanupAnySplitDetritus();
-952  fs.cleanupMergesDir();
-953}
-954
-955// Initialize split policy
-956this.splitPolicy = 
RegionSplitPolicy.create(this, conf);
-957
-958// Initialize flush policy
-959this.flushPolicy = 
FlushPolicyFactory.create(this, conf);
-960
-961long lastFlushTime = 
EnvironmentEdgeManager.currentTime();
-962for (HStore store: stores.values()) 
{
-963  
this.lastStoreFlushTimeMap.put(store, lastFlushTime);
-964}
-965
-966// Use maximum of log sequenceid or 
that which was found in stores
-967// (particularly if no recovered 
edits, seqid will be -1).
-968long maxSeqIdFromFile =
-969  
WALSplitter.getMaxRegionSequenceId(fs.getFileSystem(), fs.getRegionDir());
-970long nextSeqId = Math.max(maxSeqId, 
maxSeqIdFromFile) + 1;
-971if (writestate.writesEnabled) {
-972  
WALSplitter.writeRegionSequenceIdFile(fs.getFileSystem(), fs.getRegionDir(), 
nextSeqId - 1);
-973}
-974
-975LOG.info("Opened {}; next 
sequenceid={}", this.getRegionInfo().getShortNameToLog(), nextSeqId);
-976
-977// A region can be reopened if failed 
a split; reset flags
-978this.closing.set(false);
-979this.closed.set(false);
-980
-981if (coprocessorHost != null) {
-982  status.setStatus("Running 
coprocessor post-open hooks");
-983  coprocessorHost.postOpen();
-984}
+923LOG.debug("replaying wal for " + 
this.getRegionInfo().getEncodedName());
+924
stores.forEach(HStore::startReplayingFromWAL);
+925// Recover any edits if 
available.
+926maxSeqId = Math.max(maxSeqId,
+927  
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, 
status));
+928// Make sure mvcc is up to max.
+929this.mvcc.advanceTo(maxSeqId);
+930  } finally {
+931LOG.debug("stopping wal replay 
for " + this.getRegionInfo().getEncodedName());
+932// update the stores that we are 
done replaying
+933
stores.forEach(HStore::stopReplayingFromWAL);
+934  }
+935}
+936this.lastReplayedOpenRegionSeqId = 
maxSeqId;
+937
+938
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
+939this.writestate.flushRequested = 
false;
+940this.writestate.compacting.set(0);
+941
+942if (this.writestate.writesEnabled) 
{
+943  LOG.debug("Cleaning up temporary 
data for " + this.getRegionInfo().getEncodedName());
+944  // Remove temporary data left over 
from old regions
+945  status.setStatus("Cleaning up 
temporary data from old regions");
+946  fs.cleanupTempDir();
+947}
+948
+949if (this.writestate.writesEnabled) 
{
+950  status.setStatus("Cleaning up 
detritus from prior splits");
+951  // Get rid of any splits or merges 
that were lost in-progress.  Clean out
+952  // these directories here on open.  

[27/51] [partial] hbase-site git commit: Published site at f3d1c021de2264301f68eadb9ef126ff83d7ef53.

2018-05-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
index 7d27402..3da432b 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
@@ -2560,8 +2560,8 @@
 2552}
 2553MemStoreSize mss = 
this.memStoreSizing.getMemStoreSize();
 2554LOG.info("Flushing " + + 
storesToFlush.size() + "/" + stores.size() + " column families," +
-2555" memstore data size=" + 
StringUtils.byteDesc(mss.getDataSize()) +
-2556" memstore heap size=" + 
StringUtils.byteDesc(mss.getHeapSize()) +
+2555" dataSize=" + 
StringUtils.byteDesc(mss.getDataSize()) +
+2556" heapSize=" + 
StringUtils.byteDesc(mss.getHeapSize()) +
 2557((perCfExtras != null  
perCfExtras.length()  0)? perCfExtras.toString(): "") +
 2558((wal != null) ? "" : "; WAL is 
null, using passed sequenceid=" + sequenceId));
 2559  }
@@ -2761,10 +2761,10 @@
 2753long time = 
EnvironmentEdgeManager.currentTime() - startTime;
 2754MemStoreSize mss = 
prepareResult.totalFlushableSize.getMemStoreSize();
 2755long memstoresize = 
this.memStoreSizing.getMemStoreSize().getDataSize();
-2756String msg = "Finished memstore 
flush;"
-2757+ " data size ~" + 
StringUtils.byteDesc(mss.getDataSize()) + "/" + mss.getDataSize()
-2758+ ", heap size ~" + 
StringUtils.byteDesc(mss.getHeapSize()) + "/" + mss.getHeapSize()
-2759+ ", currentsize=" + 
StringUtils.byteDesc(memstoresize) + "/" + memstoresize
+2756String msg = "Finished flush of"
+2757+ " dataSize ~" + 
StringUtils.byteDesc(mss.getDataSize()) + "/" + mss.getDataSize()
+2758+ ", heapSize ~" + 
StringUtils.byteDesc(mss.getHeapSize()) + "/" + mss.getHeapSize()
+2759+ ", currentSize=" + 
StringUtils.byteDesc(memstoresize) + "/" + memstoresize
 2760+ " for " + 
this.getRegionInfo().getEncodedName() + " in " + time + "ms, sequenceid="
 2761+ flushOpSeqId +  ", compaction 
requested=" + compactionRequested
 2762+ ((wal == null) ? "; wal=null" 
: "");
@@ -4244,7 +4244,7 @@
 4236   * @param cellItr
 4237   * @param now
 4238   */
-4239  public void updateCellTimestamps(final 
IterableListCell cellItr, final byte[] now)
+4239  private static void 
updateCellTimestamps(final IterableListCell cellItr, final 
byte[] now)
 4240  throws IOException {
 4241for (ListCell cells: 
cellItr) {
 4242  if (cells == null) continue;
@@ -4299,12 +4299,12 @@
 4291  requestFlush();
 4292  // Don't print current limit 
because it will vary too much. The message is used as a key
 4293  // over in 
RetriesExhaustedWithDetailsException processing.
-4294  throw new 
RegionTooBusyException("Over memstore limit; regionName=" +
-4295  (this.getRegionInfo() == null? 
"unknown": this.getRegionInfo().getEncodedName()) +
-4296  ", server=" + 
(this.getRegionServerServices() == null ? "unknown":
-4297  
this.getRegionServerServices().getServerName()) +
-4298  ", blockingMemStoreSize=" +
-4299  
org.apache.hadoop.hbase.procedure2.util.StringUtils.humanSize(blockingMemStoreSize));
+4294  throw new 
RegionTooBusyException("Over memstore limit=" +
+4295
org.apache.hadoop.hbase.procedure2.util.StringUtils.humanSize(this.blockingMemStoreSize)
 +
+4296", regionName=" +
+4297  (this.getRegionInfo() == null? 
"unknown": this.getRegionInfo().getEncodedName()) +
+4298  ", server=" + 
(this.getRegionServerServices() == null? "unknown":
+4299  
this.getRegionServerServices().getServerName()));
 4300}
 4301  }
 4302
@@ -8610,7 +8610,8 @@
 8602  public void 
requestFlush(FlushLifeCycleTracker tracker) throws IOException {
 8603requestFlush0(tracker);
 8604  }
-8605}
+8605
+8606}
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
index 7d27402..3da432b 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
@@ -2560,8 +2560,8 @@
 2552}
 2553MemStoreSize mss = 
this.memStoreSizing.getMemStoreSize();
 2554

[27/51] [partial] hbase-site git commit: Published site at cf529f18a9959589fa635f78df4840472526ea2c.

2018-05-17 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
index a97dfdc..2b1b6c6 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
@@ -2370,1287 +2370,1292 @@
 2362  }
 2363
 2364  @Override
-2365  public long modifyTable(final 
TableName tableName, final TableDescriptor descriptor,
+2365  public long modifyTable(final 
TableName tableName, final TableDescriptor newDescriptor,
 2366  final long nonceGroup, final long 
nonce) throws IOException {
 2367checkInitialized();
-2368
sanityCheckTableDescriptor(descriptor);
+2368
sanityCheckTableDescriptor(newDescriptor);
 2369
 2370return 
MasterProcedureUtil.submitProcedure(
 2371new 
MasterProcedureUtil.NonceProcedureRunnable(this, nonceGroup, nonce) {
 2372  @Override
 2373  protected void run() throws 
IOException {
-2374
getMaster().getMasterCoprocessorHost().preModifyTable(tableName, descriptor);
-2375
-2376
LOG.info(getClientIdAuditPrefix() + " modify " + tableName);
+2374TableDescriptor oldDescriptor = 
getMaster().getTableDescriptors().get(tableName);
+2375
getMaster().getMasterCoprocessorHost()
+2376  .preModifyTable(tableName, 
oldDescriptor, newDescriptor);
 2377
-2378// Execute the operation 
synchronously - wait for the operation completes before continuing.
-2379//
-2380// We need to wait for the 
procedure to potentially fail due to "prepare" sanity
-2381// checks. This will block only 
the beginning of the procedure. See HBASE-19953.
-2382ProcedurePrepareLatch latch = 
ProcedurePrepareLatch.createBlockingLatch();
-2383submitProcedure(new 
ModifyTableProcedure(procedureExecutor.getEnvironment(),
-2384descriptor, latch));
-2385latch.await();
-2386
-2387
getMaster().getMasterCoprocessorHost().postModifyTable(tableName, 
descriptor);
-2388  }
-2389
-2390  @Override
-2391  protected String getDescription() 
{
-2392return "ModifyTableProcedure";
-2393  }
-2394});
-2395  }
-2396
-2397  public long restoreSnapshot(final 
SnapshotDescription snapshotDesc,
-2398  final long nonceGroup, final long 
nonce, final boolean restoreAcl) throws IOException {
-2399checkInitialized();
-2400
getSnapshotManager().checkSnapshotSupport();
-2401
-2402// Ensure namespace exists. Will 
throw exception if non-known NS.
-2403final TableName dstTable = 
TableName.valueOf(snapshotDesc.getTable());
-2404
getClusterSchema().getNamespace(dstTable.getNamespaceAsString());
-2405
-2406return 
MasterProcedureUtil.submitProcedure(
-2407new 
MasterProcedureUtil.NonceProcedureRunnable(this, nonceGroup, nonce) {
-2408  @Override
-2409  protected void run() throws 
IOException {
-2410  setProcId(
-2411
getSnapshotManager().restoreOrCloneSnapshot(snapshotDesc, getNonceKey(), 
restoreAcl));
-2412  }
-2413
-2414  @Override
-2415  protected String getDescription() 
{
-2416return 
"RestoreSnapshotProcedure";
-2417  }
-2418});
-2419  }
-2420
-2421  private void checkTableExists(final 
TableName tableName)
-2422  throws IOException, 
TableNotFoundException {
-2423if 
(!MetaTableAccessor.tableExists(getConnection(), tableName)) {
-2424  throw new 
TableNotFoundException(tableName);
-2425}
-2426  }
-2427
-2428  @Override
-2429  public void checkTableModifiable(final 
TableName tableName)
-2430  throws IOException, 
TableNotFoundException, TableNotDisabledException {
-2431if (isCatalogTable(tableName)) {
-2432  throw new IOException("Can't 
modify catalog tables");
-2433}
-2434checkTableExists(tableName);
-2435TableState ts = 
getTableStateManager().getTableState(tableName);
-2436if (!ts.isDisabled()) {
-2437  throw new 
TableNotDisabledException("Not DISABLED; " + ts);
-2438}
-2439  }
-2440
-2441  public ClusterMetrics 
getClusterMetricsWithoutCoprocessor() throws InterruptedIOException {
-2442return 
getClusterMetricsWithoutCoprocessor(EnumSet.allOf(Option.class));
-2443  }
-2444
-2445  public ClusterMetrics 
getClusterMetricsWithoutCoprocessor(EnumSetOption options)
-2446  throws InterruptedIOException {
-2447ClusterMetricsBuilder builder = 
ClusterMetricsBuilder.newBuilder();
-2448// given that hbase1 can't submit 
the request with Option,
-2449// we return all information to 
client if the list of Option is empty.
-2450if (options.isEmpty()) {
-2451  options = 
EnumSet.allOf(Option.class);
-2452}
-2453
-2454for (Option opt : options) {
-2455  switch (opt) {

[27/51] [partial] hbase-site git commit: Published site at 021f66d11d2cbb7308308093e29e69d6e7661ee9.

2018-05-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
index 4a879bb..7d27402 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
@@ -300,7 +300,7 @@
 292  private MapString, 
com.google.protobuf.Service coprocessorServiceHandlers = 
Maps.newHashMap();
 293
 294  // Track data size in all memstores
-295  private final MemStoreSizing 
memStoreSize = new MemStoreSizing();
+295  private final MemStoreSizing 
memStoreSizing = new ThreadSafeMemStoreSizing();
 296  private final RegionServicesForStores 
regionServicesForStores = new RegionServicesForStores(this);
 297
 298  // Debug possible data loss due to WAL 
off
@@ -1218,7389 +1218,7399 @@
 1210   * Increase the size of mem store in 
this region and the size of global mem
 1211   * store
 1212   */
-1213  public void 
incMemStoreSize(MemStoreSize memStoreSize) {
-1214if (this.rsAccounting != null) {
-1215  
rsAccounting.incGlobalMemStoreSize(memStoreSize);
-1216}
-1217long dataSize;
-1218synchronized (this.memStoreSize) {
-1219  
this.memStoreSize.incMemStoreSize(memStoreSize);
-1220  dataSize = 
this.memStoreSize.getDataSize();
-1221}
-1222
checkNegativeMemStoreDataSize(dataSize, memStoreSize.getDataSize());
-1223  }
-1224
-1225  public void 
decrMemStoreSize(MemStoreSize memStoreSize) {
-1226if (this.rsAccounting != null) {
-1227  
rsAccounting.decGlobalMemStoreSize(memStoreSize);
-1228}
-1229long size;
-1230synchronized (this.memStoreSize) {
-1231  
this.memStoreSize.decMemStoreSize(memStoreSize);
-1232  size = 
this.memStoreSize.getDataSize();
+1213  void incMemStoreSize(MemStoreSize mss) 
{
+1214incMemStoreSize(mss.getDataSize(), 
mss.getHeapSize(), mss.getOffHeapSize());
+1215  }
+1216
+1217  void incMemStoreSize(long 
dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) {
+1218if (this.rsAccounting != null) {
+1219  
rsAccounting.incGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, 
offHeapSizeDelta);
+1220}
+1221long dataSize =
+1222
this.memStoreSizing.incMemStoreSize(dataSizeDelta, heapSizeDelta, 
offHeapSizeDelta);
+1223
checkNegativeMemStoreDataSize(dataSize, dataSizeDelta);
+1224  }
+1225
+1226  void decrMemStoreSize(MemStoreSize 
mss) {
+1227decrMemStoreSize(mss.getDataSize(), 
mss.getHeapSize(), mss.getOffHeapSize());
+1228  }
+1229
+1230  void decrMemStoreSize(long 
dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) {
+1231if (this.rsAccounting != null) {
+1232  
rsAccounting.decGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, 
offHeapSizeDelta);
 1233}
-1234checkNegativeMemStoreDataSize(size, 
-memStoreSize.getDataSize());
-1235  }
-1236
-1237  private void 
checkNegativeMemStoreDataSize(long memStoreDataSize, long delta) {
-1238// This is extremely bad if we make 
memStoreSize negative. Log as much info on the offending
-1239// caller as possible. (memStoreSize 
might be a negative value already -- freeing memory)
-1240if (memStoreDataSize  0) {
-1241  LOG.error("Asked to modify this 
region's (" + this.toString()
-1242  + ") memStoreSize to a 
negative value which is incorrect. Current memStoreSize="
-1243  + (memStoreDataSize - delta) + 
", delta=" + delta, new Exception());
-1244}
-1245  }
-1246
-1247  @Override
-1248  public RegionInfo getRegionInfo() {
-1249return this.fs.getRegionInfo();
-1250  }
-1251
-1252  /**
-1253   * @return Instance of {@link 
RegionServerServices} used by this HRegion.
-1254   * Can be null.
-1255   */
-1256  RegionServerServices 
getRegionServerServices() {
-1257return this.rsServices;
-1258  }
-1259
-1260  @Override
-1261  public long getReadRequestsCount() {
-1262return readRequestsCount.sum();
-1263  }
-1264
-1265  @Override
-1266  public long 
getFilteredReadRequestsCount() {
-1267return 
filteredReadRequestsCount.sum();
-1268  }
-1269
-1270  @Override
-1271  public long getWriteRequestsCount() 
{
-1272return writeRequestsCount.sum();
-1273  }
-1274
-1275  @Override
-1276  public long getMemStoreDataSize() {
-1277return memStoreSize.getDataSize();
-1278  }
-1279
-1280  @Override
-1281  public long getMemStoreHeapSize() {
-1282return memStoreSize.getHeapSize();
-1283  }
-1284
-1285  @Override
-1286  public long getMemStoreOffHeapSize() 
{
-1287return 
memStoreSize.getOffHeapSize();
-1288  }
-1289
-1290  /** @return store services for this 
region, to access services required by store level 

[27/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.

2018-05-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
index 2510283..418c60c 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html
@@ -77,77 +77,77 @@
 069import 
org.apache.hadoop.hbase.client.RowMutations;
 070import 
org.apache.hadoop.hbase.client.Scan;
 071import 
org.apache.hadoop.hbase.client.Table;
-072import 
org.apache.hadoop.hbase.filter.BinaryComparator;
-073import 
org.apache.hadoop.hbase.filter.Filter;
-074import 
org.apache.hadoop.hbase.filter.FilterAllFilter;
-075import 
org.apache.hadoop.hbase.filter.FilterList;
-076import 
org.apache.hadoop.hbase.filter.PageFilter;
-077import 
org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
-078import 
org.apache.hadoop.hbase.filter.WhileMatchFilter;
-079import 
org.apache.hadoop.hbase.io.compress.Compression;
-080import 
org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
-081import 
org.apache.hadoop.hbase.io.hfile.RandomDistribution;
-082import 
org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
-083import 
org.apache.hadoop.hbase.regionserver.BloomType;
-084import 
org.apache.hadoop.hbase.regionserver.CompactingMemStore;
-085import 
org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration;
-086import 
org.apache.hadoop.hbase.trace.SpanReceiverHost;
-087import 
org.apache.hadoop.hbase.trace.TraceUtil;
-088import 
org.apache.hadoop.hbase.util.ByteArrayHashKey;
-089import 
org.apache.hadoop.hbase.util.Bytes;
-090import 
org.apache.hadoop.hbase.util.Hash;
-091import 
org.apache.hadoop.hbase.util.MurmurHash;
-092import 
org.apache.hadoop.hbase.util.Pair;
-093import 
org.apache.hadoop.hbase.util.YammerHistogramUtils;
-094import 
org.apache.hadoop.io.LongWritable;
-095import org.apache.hadoop.io.Text;
-096import org.apache.hadoop.mapreduce.Job;
-097import 
org.apache.hadoop.mapreduce.Mapper;
-098import 
org.apache.hadoop.mapreduce.lib.input.NLineInputFormat;
-099import 
org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
-100import 
org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer;
-101import org.apache.hadoop.util.Tool;
-102import 
org.apache.hadoop.util.ToolRunner;
-103import 
org.apache.htrace.core.ProbabilitySampler;
-104import org.apache.htrace.core.Sampler;
-105import 
org.apache.htrace.core.TraceScope;
-106import 
org.apache.yetus.audience.InterfaceAudience;
-107import org.slf4j.Logger;
-108import org.slf4j.LoggerFactory;
-109import 
org.apache.hbase.thirdparty.com.google.common.base.MoreObjects;
-110import 
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-111
-112/**
-113 * Script used evaluating HBase 
performance and scalability.  Runs a HBase
-114 * client that steps through one of a set 
of hardcoded tests or 'experiments'
-115 * (e.g. a random reads test, a random 
writes test, etc.). Pass on the
-116 * command-line which test to run and how 
many clients are participating in
-117 * this experiment. Run {@code 
PerformanceEvaluation --help} to obtain usage.
-118 *
-119 * pThis class sets up and runs 
the evaluation programs described in
-120 * Section 7, iPerformance 
Evaluation/i, of the a
-121 * 
href="http://labs.google.com/papers/bigtable.html"Bigtable/a;
-122 * paper, pages 8-10.
-123 *
-124 * pBy default, runs as a 
mapreduce job where each mapper runs a single test
-125 * client. Can also run as a 
non-mapreduce, multithreaded application by
-126 * specifying {@code --nomapred}. Each 
client does about 1GB of data, unless
-127 * specified otherwise.
-128 */
-129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
-130public class PerformanceEvaluation 
extends Configured implements Tool {
-131  static final String RANDOM_SEEK_SCAN = 
"randomSeekScan";
-132  static final String RANDOM_READ = 
"randomRead";
-133  private static final Logger LOG = 
LoggerFactory.getLogger(PerformanceEvaluation.class.getName());
-134  private static final ObjectMapper 
MAPPER = new ObjectMapper();
-135  static {
-136
MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true);
-137  }
-138
-139  public static final String TABLE_NAME = 
"TestTable";
-140  public static final byte[] FAMILY_NAME 
= Bytes.toBytes("info");
-141  public static final byte [] COLUMN_ZERO 
= Bytes.toBytes("" + 0);
-142  public static final byte [] 
QUALIFIER_NAME = COLUMN_ZERO;
+072import 
org.apache.hadoop.hbase.client.metrics.ScanMetrics;
+073import 
org.apache.hadoop.hbase.filter.BinaryComparator;
+074import 

[27/51] [partial] hbase-site git commit: Published site at 87f5b5f3411d96c31b4cb61b9a57ced22be91d1f.

2018-05-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html
index e1bc325..63e7421 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html
@@ -66,5125 +66,5224 @@
 058import 
java.util.concurrent.TimeoutException;
 059import 
java.util.concurrent.atomic.AtomicBoolean;
 060import 
java.util.concurrent.atomic.AtomicInteger;
-061import org.apache.commons.io.IOUtils;
-062import 
org.apache.commons.lang3.RandomStringUtils;
-063import 
org.apache.commons.lang3.StringUtils;
-064import 
org.apache.hadoop.conf.Configuration;
-065import 
org.apache.hadoop.conf.Configured;
-066import 
org.apache.hadoop.fs.FSDataOutputStream;
-067import org.apache.hadoop.fs.FileStatus;
-068import org.apache.hadoop.fs.FileSystem;
-069import org.apache.hadoop.fs.Path;
-070import 
org.apache.hadoop.fs.permission.FsAction;
-071import 
org.apache.hadoop.fs.permission.FsPermission;
-072import 
org.apache.hadoop.hbase.Abortable;
-073import org.apache.hadoop.hbase.Cell;
-074import 
org.apache.hadoop.hbase.CellUtil;
-075import 
org.apache.hadoop.hbase.ClusterMetrics;
-076import 
org.apache.hadoop.hbase.ClusterMetrics.Option;
-077import 
org.apache.hadoop.hbase.HBaseConfiguration;
-078import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-079import 
org.apache.hadoop.hbase.HConstants;
-080import 
org.apache.hadoop.hbase.HRegionInfo;
-081import 
org.apache.hadoop.hbase.HRegionLocation;
-082import 
org.apache.hadoop.hbase.KeyValue;
-083import 
org.apache.hadoop.hbase.MasterNotRunningException;
-084import 
org.apache.hadoop.hbase.MetaTableAccessor;
-085import 
org.apache.hadoop.hbase.RegionLocations;
-086import 
org.apache.hadoop.hbase.ServerName;
-087import 
org.apache.hadoop.hbase.TableName;
-088import 
org.apache.hadoop.hbase.ZooKeeperConnectionException;
-089import 
org.apache.hadoop.hbase.client.Admin;
-090import 
org.apache.hadoop.hbase.client.ClusterConnection;
-091import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
-092import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
-093import 
org.apache.hadoop.hbase.client.Connection;
-094import 
org.apache.hadoop.hbase.client.ConnectionFactory;
-095import 
org.apache.hadoop.hbase.client.Delete;
-096import 
org.apache.hadoop.hbase.client.Get;
-097import 
org.apache.hadoop.hbase.client.Put;
-098import 
org.apache.hadoop.hbase.client.RegionInfo;
-099import 
org.apache.hadoop.hbase.client.RegionInfoBuilder;
-100import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-101import 
org.apache.hadoop.hbase.client.Result;
-102import 
org.apache.hadoop.hbase.client.RowMutations;
-103import 
org.apache.hadoop.hbase.client.Table;
-104import 
org.apache.hadoop.hbase.client.TableDescriptor;
-105import 
org.apache.hadoop.hbase.client.TableDescriptorBuilder;
-106import 
org.apache.hadoop.hbase.client.TableState;
-107import 
org.apache.hadoop.hbase.io.FileLink;
-108import 
org.apache.hadoop.hbase.io.HFileLink;
-109import 
org.apache.hadoop.hbase.io.hfile.CacheConfig;
-110import 
org.apache.hadoop.hbase.io.hfile.HFile;
-111import 
org.apache.hadoop.hbase.log.HBaseMarkers;
-112import 
org.apache.hadoop.hbase.master.MasterFileSystem;
-113import 
org.apache.hadoop.hbase.master.RegionState;
-114import 
org.apache.hadoop.hbase.regionserver.HRegion;
-115import 
org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
-116import 
org.apache.hadoop.hbase.regionserver.StoreFileInfo;
-117import 
org.apache.hadoop.hbase.replication.ReplicationException;
-118import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-119import 
org.apache.hadoop.hbase.security.UserProvider;
-120import 
org.apache.hadoop.hbase.util.Bytes.ByteArrayComparator;
-121import 
org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE;
-122import 
org.apache.hadoop.hbase.util.hbck.HFileCorruptionChecker;
-123import 
org.apache.hadoop.hbase.util.hbck.ReplicationChecker;
-124import 
org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandler;
-125import 
org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandlerImpl;
-126import org.apache.hadoop.hbase.wal.WAL;
-127import 
org.apache.hadoop.hbase.wal.WALFactory;
-128import 
org.apache.hadoop.hbase.wal.WALSplitter;
-129import 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator;
-130import 
org.apache.hadoop.hbase.zookeeper.ZKUtil;
-131import 
org.apache.hadoop.hbase.zookeeper.ZKWatcher;
-132import 
org.apache.hadoop.hbase.zookeeper.ZNodePaths;
-133import 
org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException;
-134import 
org.apache.hadoop.ipc.RemoteException;
-135import 
org.apache.hadoop.security.UserGroupInformation;
-136import 

[27/51] [partial] hbase-site git commit: Published site at 2912c953551bedbfbf30c32c156ed7bb187d54c3.

2018-04-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html
index 56ec3ac..8f4cacb 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html
@@ -229,9 +229,9 @@
 221   * Note that we don't document 
--expected-count, because it's intended for test.
 222   */
 223  private static void printUsage() {
-224System.err.println("Usage: RowCounter 
[options] tablename " +
-225"[--starttime=[start] 
--endtime=[end] " +
-226
"[--range=[startKey],[endKey][;[startKey],[endKey]...]] [column1 
column2...]");
+224System.err.println("Usage: hbase 
rowcounter [options] tablename "
+225+ "[--starttime=start 
--endtime=end] "
+226+ 
"[--range=[startKey],[endKey][;[startKey],[endKey]...]] [column1 
column2...]");
 227System.err.println("For performance 
consider the following options:\n"
 228+ 
"-Dhbase.client.scanner.caching=100\n"
 229+ 
"-Dmapreduce.map.speculative=false");

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
index a5477ac..9501e97 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html
@@ -357,46 +357,51 @@
 349  System.err.println("ERROR: " + 
errorMsg);
 350}
 351System.err.println("Usage: " + NAME + 
" [options] wal inputdir tables [tableMappings]");
-352System.err.println("Read all WAL 
entries for tables.");
-353System.err.println("If no tables 
(\"\") are specific, all tables are imported.");
-354System.err.println("(Careful, even 
hbase:meta entries will be imported" + " in that case.)");
-355System.err.println("Otherwise 
tables is a comma separated list of tables.\n");
-356System.err.println("The WAL entries 
can be mapped to new set of tables via tableMapping.");
-357
System.err.println("tableMapping is a command separated list of 
targettables.");
+352System.err.println("Replay all WAL 
files into HBase.");
+353System.err.println("tables is 
a comma separated list of tables.");
+354System.err.println("If no tables 
(\"\") are specified, all tables are imported.");
+355System.err.println("(Be careful, 
hbase:meta entries will be imported in this case.)\n");
+356System.err.println("WAL entries can 
be mapped to new set of tables via tableMappings.");
+357
System.err.println("tableMappings is a comma separated list of target 
tables.");
 358System.err.println("If specified, 
each table in tables must have a mapping.\n");
 359System.err.println("By default " + 
NAME + " will load data directly into HBase.");
-360System.err.println("To generate 
HFiles for a bulk data load instead, pass the option:");
+360System.err.println("To generate 
HFiles for a bulk data load instead, pass the following option:");
 361System.err.println("  -D" + 
BULK_OUTPUT_CONF_KEY + "=/path/for/output");
 362System.err.println("  (Only one table 
can be specified, and no mapping is allowed!)");
-363System.err.println("Other options: 
(specify time range to WAL edit to consider)");
+363System.err.println("Time range 
options:");
 364System.err.println("  -D" + 
WALInputFormat.START_TIME_KEY + "=[date|ms]");
 365System.err.println("  -D" + 
WALInputFormat.END_TIME_KEY + "=[date|ms]");
-366System.err.println("   -D " + 
JOB_NAME_CONF_KEY
-367+ "=jobName - use the specified 
mapreduce job name for the wal player");
-368System.err.println("For performance 
also consider the following options:\n"
-369+ "  
-Dmapreduce.map.speculative=false\n" + "  
-Dmapreduce.reduce.speculative=false");
-370  }
-371
-372  /**
-373   * Main entry point.
-374   * @param args The command line 
parameters.
-375   * @throws Exception When running the 
job fails.
-376   */
-377  public static void main(String[] args) 
throws Exception {
-378int ret = ToolRunner.run(new 
WALPlayer(HBaseConfiguration.create()), args);
-379System.exit(ret);
-380  }
-381
-382  @Override
-383  public int run(String[] args) throws 
Exception {
-384if (args.length  2) {
-385  usage("Wrong number of arguments: " 
+ args.length);
-386  System.exit(-1);
-387}
-388Job job = 

[27/51] [partial] hbase-site git commit: Published site at 2a2258656b2fcd92b967131b6c1f037363553bc4.

2018-03-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html 
b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html
index efc1791..6e81af2 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class RegionMover.MoveWithoutAck
+private static class RegionMover.MoveWithoutAck
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true;
 title="class or interface in java.util.concurrent">Callablehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
 Move Regions without Acknowledging.Usefule in case of RS 
shutdown as we might want to shut the
@@ -228,7 +228,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
 
 
 admin
-privateAdmin admin
+privateAdmin admin
 
 
 
@@ -237,7 +237,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
 
 
 region
-privateRegionInfo region
+privateRegionInfo region
 
 
 
@@ -246,7 +246,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
 
 
 targetServer
-privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String targetServer
+privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String targetServer
 
 
 
@@ -255,7 +255,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
 
 
 movedRegions
-privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo movedRegions
+privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo movedRegions
 
 
 
@@ -264,7 +264,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
 
 
 sourceServer
-privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String sourceServer
+privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String sourceServer
 
 
 
@@ -281,7 +281,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
 
 
 MoveWithoutAck
-publicMoveWithoutAck(Adminadmin,
+publicMoveWithoutAck(Adminadmin,
   RegionInforegionInfo,
   https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringsourceServer,
   https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtargetServer,
@@ -302,7 +302,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre
 
 
 call
-publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Booleancall()
+publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Booleancall()
 
 Specified by:
 https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true#call--;
 title="class or interface in java.util.concurrent">callin 
interfacehttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true;
 title="class or interface in java.util.concurrent">Callablehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html 
b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html
index c451701..d4d8db5 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public static class RegionMover.RegionMoverBuilder
+public static class 

[27/51] [partial] hbase-site git commit: Published site at e468b4022f76688851b3e0c34722f01a56bd624f.

2018-03-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html 
b/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
index dbb2f3e..c887b35 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
@@ -36,843 +36,849 @@
 028import 
org.apache.hadoop.hbase.CompareOperator;
 029import 
org.apache.hadoop.hbase.HTableDescriptor;
 030import 
org.apache.hadoop.hbase.TableName;
-031import 
org.apache.yetus.audience.InterfaceAudience;
-032import 
org.apache.hadoop.hbase.client.coprocessor.Batch;
-033import 
org.apache.hadoop.hbase.filter.CompareFilter;
-034import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
-035
-036import com.google.protobuf.Descriptors;
-037import com.google.protobuf.Message;
-038import com.google.protobuf.Service;
-039import 
com.google.protobuf.ServiceException;
-040
-041/**
-042 * Used to communicate with a single 
HBase table.
-043 * Obtain an instance from a {@link 
Connection} and call {@link #close()} afterwards.
-044 *
-045 * pTable can be used to get, 
put, delete or scan data from a table.
-046 * @see ConnectionFactory
-047 * @see Connection
-048 * @see Admin
-049 * @see RegionLocator
-050 * @since 0.99.0
-051 */
-052@InterfaceAudience.Public
-053public interface Table extends Closeable 
{
-054  /**
-055   * Gets the fully qualified table name 
instance of this table.
-056   */
-057  TableName getName();
-058
-059  /**
-060   * Returns the {@link 
org.apache.hadoop.conf.Configuration} object used by this instance.
-061   * p
-062   * The reference returned is not a 
copy, so any change made to it will
-063   * affect this instance.
-064   */
-065  Configuration getConfiguration();
-066
-067  /**
-068   * Gets the {@link 
org.apache.hadoop.hbase.HTableDescriptor table descriptor} for this table.
-069   * @throws java.io.IOException if a 
remote or network exception occurs.
-070   * @deprecated since 2.0 version and 
will be removed in 3.0 version.
-071   * use {@link 
#getDescriptor()}
-072   */
-073  @Deprecated
-074  HTableDescriptor getTableDescriptor() 
throws IOException;
-075
-076  /**
-077   * Gets the {@link 
org.apache.hadoop.hbase.client.TableDescriptor table descriptor} for this 
table.
-078   * @throws java.io.IOException if a 
remote or network exception occurs.
-079   */
-080  TableDescriptor getDescriptor() throws 
IOException;
-081
-082  /**
-083   * Test for the existence of columns in 
the table, as specified by the Get.
-084   * p
-085   *
-086   * This will return true if the Get 
matches one or more keys, false if not.
-087   * p
-088   *
-089   * This is a server-side call so it 
prevents any data from being transfered to
-090   * the client.
-091   *
-092   * @param get the Get
-093   * @return true if the specified Get 
matches one or more keys, false if not
-094   * @throws IOException e
-095   */
-096  boolean exists(Get get) throws 
IOException;
-097
-098  /**
-099   * Test for the existence of columns in 
the table, as specified by the Gets.
-100   * p
-101   *
-102   * This will return an array of 
booleans. Each value will be true if the related Get matches
-103   * one or more keys, false if not.
-104   * p
-105   *
-106   * This is a server-side call so it 
prevents any data from being transferred to
-107   * the client.
-108   *
-109   * @param gets the Gets
-110   * @return Array of boolean.  True if 
the specified Get matches one or more keys, false if not.
-111   * @throws IOException e
-112   */
-113  boolean[] exists(ListGet gets) 
throws IOException;
-114
-115  /**
-116   * Test for the existence of columns in 
the table, as specified by the Gets.
-117   * This will return an array of 
booleans. Each value will be true if the related Get matches
-118   * one or more keys, false if not.
-119   * This is a server-side call so it 
prevents any data from being transferred to
-120   * the client.
-121   *
-122   * @param gets the Gets
-123   * @return Array of boolean.  True if 
the specified Get matches one or more keys, false if not.
-124   * @throws IOException e
-125   * @deprecated since 2.0 version and 
will be removed in 3.0 version.
-126   * use {@link 
#exists(List)}
-127   */
-128  @Deprecated
-129  default boolean[] 
existsAll(ListGet gets) throws IOException {
-130return exists(gets);
-131  }
-132
-133  /**
-134   * Method that does a batch call on 
Deletes, Gets, Puts, Increments, Appends, RowMutations.
-135   * The ordering of execution of the 
actions is not defined. Meaning if you do a Put and a
-136   * Get in the same {@link #batch} call, 
you will not necessarily be
-137   * guaranteed that the Get returns what 
the Put had put.
-138   *
-139   * @param actions list of Get, Put, 
Delete, Increment, Append, 

[27/51] [partial] hbase-site git commit: Published site at 64061f896fe21512504e3886a400759e88b519da.

2018-03-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
index 5360e0a..d4296e6 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
@@ -627,13 +627,18 @@ service.
https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
 title="class or interface in 
java.util.concurrent">ExecutorServicepool)
 
 
+static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListPairhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long
+MetaTableAccessor.getTableEncodedRegionNameAndLastBarrier(Connectionconn,
+   TableNametableName)
+
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
 MetaTableAccessor.getTableRegions(Connectionconnection,
TableNametableName)
 Gets all of the regions of the specified table.
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
 MetaTableAccessor.getTableRegions(Connectionconnection,
TableNametableName,
@@ -641,14 +646,14 @@ service.
 Gets all of the regions of the specified table.
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListPairRegionInfo,ServerName
 MetaTableAccessor.getTableRegionsAndLocations(Connectionconnection,
TableNametableName)
 Do not use this method to get meta table regions, use 
methods in MetaTableLocator instead.
 
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListPairRegionInfo,ServerName
 MetaTableAccessor.getTableRegionsAndLocations(Connectionconnection,
TableNametableName,
@@ -656,37 +661,37 @@ service.
 Do not use this method to get meta table regions, use 
methods in MetaTableLocator instead.
 
 
-
+
 static byte[]
 MetaTableAccessor.getTableStartRowForMeta(TableNametableName,
MetaTableAccessor.QueryTypetype)
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalTableState
 AsyncMetaTableAccessor.getTableState(AsyncTable?metaTable,
  TableNametableName)
 
-
+
 static TableState
 MetaTableAccessor.getTableState(Connectionconn,
  TableNametableName)
 Fetch table state for given table from META table
 
 
-
+
 static byte[]
 MetaTableAccessor.getTableStopRowForMeta(TableNametableName,
   MetaTableAccessor.QueryTypetype)
 
-
+
 static boolean
 TableName.isMetaTableName(TableNametn)
 
-
+
 TableDescriptor
 TableDescriptors.remove(TableNametablename)
 
-
+
 static void
 MetaTableAccessor.scanMeta(Connectionconnection,
 MetaTableAccessor.Visitorvisitor,
@@ -697,7 +702,7 @@ service.
  given row.
 
 
-
+
 static void
 MetaTableAccessor.scanMeta(Connectionconnection,
 TableNametable,
@@ -705,25 +710,25 @@ service.
 intmaxRows,
 MetaTableAccessor.Visitorvisitor)
 
-
+
 static void
 MetaTableAccessor.scanMetaForTableRegions(Connectionconnection,
MetaTableAccessor.Visitorvisitor,
TableNametableName)
 
-
+
 static https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
 AsyncMetaTableAccessor.tableExists(AsyncTable?metaTable,
TableNametableName)
 
-
+
 static boolean
 MetaTableAccessor.tableExists(Connectionconnection,
TableNametableName)
 Checks if the specified table exists.
 
 
-
+
 static void
 MetaTableAccessor.updateTableState(Connectionconn,
 TableNametableName,
@@ -2052,119 +2057,119 @@ service.
 
 
 private TableName
-SnapshotDescription.table
+RegionCoprocessorRpcChannel.table
 
 
 private TableName

[27/51] [partial] hbase-site git commit: Published site at 4cb40e6d846ce1f28ffb40d388c9efb753197813.

2018-03-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/class-use/Size.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Size.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/Size.html
index 3f1b032..e8d1010 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/Size.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/Size.html
@@ -189,130 +189,130 @@
 
 
 Size
-RegionLoad.getBloomFilterSize()
-Deprecated.
-
-
-
-Size
 RegionMetrics.getBloomFilterSize()
 
-
-Size
-RegionMetricsBuilder.RegionMetricsImpl.getBloomFilterSize()
-
 
 Size
-ServerLoad.getMaxHeapSize()
+RegionLoad.getBloomFilterSize()
 Deprecated.
 
 
 
 Size
-ServerMetrics.getMaxHeapSize()
+RegionMetricsBuilder.RegionMetricsImpl.getBloomFilterSize()
 
 
 Size
-ServerMetricsBuilder.ServerMetricsImpl.getMaxHeapSize()
+ServerMetrics.getMaxHeapSize()
 
 
 Size
-RegionLoad.getMemStoreSize()
+ServerLoad.getMaxHeapSize()
 Deprecated.
 
 
 
 Size
-RegionMetrics.getMemStoreSize()
+ServerMetricsBuilder.ServerMetricsImpl.getMaxHeapSize()
 
 
 Size
-RegionMetricsBuilder.RegionMetricsImpl.getMemStoreSize()
+RegionMetrics.getMemStoreSize()
 
 
 Size
-RegionLoad.getStoreFileIndexSize()
+RegionLoad.getMemStoreSize()
 Deprecated.
 
 
 
 Size
+RegionMetricsBuilder.RegionMetricsImpl.getMemStoreSize()
+
+
+Size
 RegionMetrics.getStoreFileIndexSize()
 TODO: why we pass the same value to different counters? 
Currently, the value from
  getStoreFileIndexSize() is same with getStoreFileRootLevelIndexSize()
  see HRegionServer#createRegionLoad.
 
 
-
-Size
-RegionMetricsBuilder.RegionMetricsImpl.getStoreFileIndexSize()
-
 
 Size
-RegionLoad.getStoreFileRootLevelIndexSize()
+RegionLoad.getStoreFileIndexSize()
 Deprecated.
 
 
 
 Size
-RegionMetrics.getStoreFileRootLevelIndexSize()
+RegionMetricsBuilder.RegionMetricsImpl.getStoreFileIndexSize()
 
 
 Size
-RegionMetricsBuilder.RegionMetricsImpl.getStoreFileRootLevelIndexSize()
+RegionMetrics.getStoreFileRootLevelIndexSize()
 
 
 Size
-RegionLoad.getStoreFileSize()
+RegionLoad.getStoreFileRootLevelIndexSize()
 Deprecated.
 
 
 
 Size
-RegionMetrics.getStoreFileSize()
+RegionMetricsBuilder.RegionMetricsImpl.getStoreFileRootLevelIndexSize()
 
 
 Size
-RegionMetricsBuilder.RegionMetricsImpl.getStoreFileSize()
+RegionMetrics.getStoreFileSize()
 
 
 Size
-RegionLoad.getStoreFileUncompressedDataIndexSize()
+RegionLoad.getStoreFileSize()
 Deprecated.
 
 
 
 Size
-RegionMetrics.getStoreFileUncompressedDataIndexSize()
+RegionMetricsBuilder.RegionMetricsImpl.getStoreFileSize()
 
 
 Size
-RegionMetricsBuilder.RegionMetricsImpl.getStoreFileUncompressedDataIndexSize()
+RegionMetrics.getStoreFileUncompressedDataIndexSize()
 
 
 Size
-RegionLoad.getUncompressedStoreFileSize()
+RegionLoad.getStoreFileUncompressedDataIndexSize()
 Deprecated.
 
 
 
 Size
-RegionMetrics.getUncompressedStoreFileSize()
+RegionMetricsBuilder.RegionMetricsImpl.getStoreFileUncompressedDataIndexSize()
 
 
 Size
-RegionMetricsBuilder.RegionMetricsImpl.getUncompressedStoreFileSize()
+RegionMetrics.getUncompressedStoreFileSize()
 
 
 Size
-ServerLoad.getUsedHeapSize()
+RegionLoad.getUncompressedStoreFileSize()
 Deprecated.
 
 
 
 Size
+RegionMetricsBuilder.RegionMetricsImpl.getUncompressedStoreFileSize()
+
+
+Size
 ServerMetrics.getUsedHeapSize()
 
+
+Size
+ServerLoad.getUsedHeapSize()
+Deprecated.
+
+
 
 Size
 ServerMetricsBuilder.ServerMetricsImpl.getUsedHeapSize()

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html
index 63833f7..72d579d 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html
@@ -122,11 +122,11 @@
 
 
 TableDescriptors
-HMaster.getTableDescriptors()
+MasterServices.getTableDescriptors()
 
 
 TableDescriptors
-MasterServices.getTableDescriptors()
+HMaster.getTableDescriptors()
 
 
 



[27/51] [partial] hbase-site git commit: Published site at 8ab7b20f48951d77945181024f5e15842bc253c4.

2018-03-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
index ecf500c..0cd5a4e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
@@ -238,8355 +238,8368 @@
 230  public static final String 
HBASE_MAX_CELL_SIZE_KEY = "hbase.server.keyvalue.maxsize";
 231  public static final int 
DEFAULT_MAX_CELL_SIZE = 10485760;
 232
-233  public static final String 
HBASE_REGIONSERVER_MINIBATCH_SIZE =
-234  
"hbase.regionserver.minibatch.size";
-235  public static final int 
DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2;
-236
-237  /**
-238   * This is the global default value for 
durability. All tables/mutations not
-239   * defining a durability or using 
USE_DEFAULT will default to this value.
-240   */
-241  private static final Durability 
DEFAULT_DURABILITY = Durability.SYNC_WAL;
+233  /**
+234   * This is the global default value for 
durability. All tables/mutations not
+235   * defining a durability or using 
USE_DEFAULT will default to this value.
+236   */
+237  private static final Durability 
DEFAULT_DURABILITY = Durability.SYNC_WAL;
+238
+239  public static final String 
HBASE_REGIONSERVER_MINIBATCH_SIZE =
+240  
"hbase.regionserver.minibatch.size";
+241  public static final int 
DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2;
 242
-243  final AtomicBoolean closed = new 
AtomicBoolean(false);
-244
-245  /* Closing can take some time; use the 
closing flag if there is stuff we don't
-246   * want to do while in closing state; 
e.g. like offer this region up to the
-247   * master as a region to close if the 
carrying regionserver is overloaded.
-248   * Once set, it is never cleared.
-249   */
-250  final AtomicBoolean closing = new 
AtomicBoolean(false);
-251
-252  /**
-253   * The max sequence id of flushed data 
on this region. There is no edit in memory that is
-254   * less that this sequence id.
-255   */
-256  private volatile long maxFlushedSeqId = 
HConstants.NO_SEQNUM;
-257
-258  /**
-259   * Record the sequence id of last flush 
operation. Can be in advance of
-260   * {@link #maxFlushedSeqId} when 
flushing a single column family. In this case,
-261   * {@link #maxFlushedSeqId} will be 
older than the oldest edit in memory.
-262   */
-263  private volatile long lastFlushOpSeqId 
= HConstants.NO_SEQNUM;
-264
-265  /**
-266   * The sequence id of the last replayed 
open region event from the primary region. This is used
-267   * to skip entries before this due to 
the possibility of replay edits coming out of order from
-268   * replication.
-269   */
-270  protected volatile long 
lastReplayedOpenRegionSeqId = -1L;
-271  protected volatile long 
lastReplayedCompactionSeqId = -1L;
-272
-273  
//
-274  // Members
-275  
//
-276
-277  // map from a locked row to the context 
for that lock including:
-278  // - CountDownLatch for threads waiting 
on that row
-279  // - the thread that owns the lock 
(allow reentrancy)
-280  // - reference count of (reentrant) 
locks held by the thread
-281  // - the row itself
-282  private final 
ConcurrentHashMapHashedBytes, RowLockContext lockedRows =
-283  new ConcurrentHashMap();
-284
-285  protected final Mapbyte[], 
HStore stores =
-286  new 
ConcurrentSkipListMap(Bytes.BYTES_RAWCOMPARATOR);
+243  public static final String 
WAL_HSYNC_CONF_KEY = "hbase.wal.hsync";
+244  public static final boolean 
DEFAULT_WAL_HSYNC = false;
+245
+246  final AtomicBoolean closed = new 
AtomicBoolean(false);
+247
+248  /* Closing can take some time; use the 
closing flag if there is stuff we don't
+249   * want to do while in closing state; 
e.g. like offer this region up to the
+250   * master as a region to close if the 
carrying regionserver is overloaded.
+251   * Once set, it is never cleared.
+252   */
+253  final AtomicBoolean closing = new 
AtomicBoolean(false);
+254
+255  /**
+256   * The max sequence id of flushed data 
on this region. There is no edit in memory that is
+257   * less that this sequence id.
+258   */
+259  private volatile long maxFlushedSeqId = 
HConstants.NO_SEQNUM;
+260
+261  /**
+262   * Record the sequence id of last flush 
operation. Can be in advance of
+263   * {@link #maxFlushedSeqId} when 
flushing a single column family. In this case,
+264   * {@link #maxFlushedSeqId} will be 
older than the oldest edit in memory.
+265   */
+266  private volatile 

[27/51] [partial] hbase-site git commit: Published site at 00095a2ef9442e3fd86c04876c9d91f2f8b23ad8.

2018-03-17 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
index f47d627..c3d225c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html
@@ -117,219 +117,219 @@
 109   */
 110  public static boolean 
archiveRegion(FileSystem fs, Path rootdir, Path tableDir, Path regionDir)
 111  throws IOException {
-112if (LOG.isDebugEnabled()) {
-113  LOG.debug("ARCHIVING " + 
regionDir.toString());
-114}
-115
-116// otherwise, we archive the files
-117// make sure we can archive
-118if (tableDir == null || regionDir == 
null) {
-119  LOG.error("No archive directory 
could be found because tabledir (" + tableDir
-120  + ") or regiondir (" + 
regionDir + "was null. Deleting files instead.");
-121  deleteRegionWithoutArchiving(fs, 
regionDir);
-122  // we should have archived, but 
failed to. Doesn't matter if we deleted
-123  // the archived files correctly or 
not.
-124  return false;
-125}
-126
-127// make sure the regiondir lives 
under the tabledir
-128
Preconditions.checkArgument(regionDir.toString().startsWith(tableDir.toString()));
-129Path regionArchiveDir = 
HFileArchiveUtil.getRegionArchiveDir(rootdir,
-130FSUtils.getTableName(tableDir),
-131regionDir.getName());
-132
-133FileStatusConverter getAsFile = new 
FileStatusConverter(fs);
-134// otherwise, we attempt to archive 
the store files
-135
-136// build collection of just the store 
directories to archive
-137CollectionFile toArchive = 
new ArrayList();
-138final PathFilter dirFilter = new 
FSUtils.DirFilter(fs);
-139PathFilter nonHidden = new 
PathFilter() {
-140  @Override
-141  public boolean accept(Path file) 
{
-142return dirFilter.accept(file) 
 !file.getName().toString().startsWith(".");
-143  }
-144};
-145FileStatus[] storeDirs = 
FSUtils.listStatus(fs, regionDir, nonHidden);
-146// if there no files, we can just 
delete the directory and return;
-147if (storeDirs == null) {
-148  LOG.debug("Region directory " + 
regionDir + " empty.");
-149  return 
deleteRegionWithoutArchiving(fs, regionDir);
-150}
-151
-152// convert the files in the region to 
a File
-153
toArchive.addAll(Lists.transform(Arrays.asList(storeDirs), getAsFile));
-154LOG.debug("Archiving " + 
toArchive);
-155ListFile failedArchive = 
resolveAndArchive(fs, regionArchiveDir, toArchive,
-156
EnvironmentEdgeManager.currentTime());
-157if (!failedArchive.isEmpty()) {
-158  throw new 
FailedArchiveException("Failed to archive/delete all the files for region:"
-159  + regionDir.getName() + " into 
" + regionArchiveDir
-160  + ". Something is probably awry 
on the filesystem.",
-161  
Collections2.transform(failedArchive, FUNC_FILE_TO_PATH));
-162}
-163// if that was successful, then we 
delete the region
-164return 
deleteRegionWithoutArchiving(fs, regionDir);
-165  }
-166
-167  /**
-168   * Remove from the specified region the 
store files of the specified column family,
-169   * either by archiving them or outright 
deletion
-170   * @param fs the filesystem where the 
store files live
-171   * @param conf {@link Configuration} to 
examine to determine the archive directory
-172   * @param parent Parent region hosting 
the store files
-173   * @param tableDir {@link Path} to 
where the table is being stored (for building the archive path)
-174   * @param family the family hosting the 
store files
-175   * @throws IOException if the files 
could not be correctly disposed.
-176   */
-177  public static void 
archiveFamily(FileSystem fs, Configuration conf,
-178  RegionInfo parent, Path tableDir, 
byte[] family) throws IOException {
-179Path familyDir = new Path(tableDir, 
new Path(parent.getEncodedName(), Bytes.toString(family)));
-180archiveFamilyByFamilyDir(fs, conf, 
parent, familyDir, family);
-181  }
-182
-183  /**
-184   * Removes from the specified region 
the store files of the specified column family,
-185   * either by archiving them or outright 
deletion
-186   * @param fs the filesystem where the 
store files live
-187   * @param conf {@link Configuration} to 
examine to determine the archive directory
-188   * @param parent Parent region hosting 
the store files
-189   * @param familyDir {@link Path} to 
where the family is being stored
-190   * @param family the family hosting the 
store files
-191   * @throws IOException if the files 
could not be correctly disposed.
-192   */

[27/51] [partial] hbase-site git commit: Published site at 22f4def942f8a3367d0ca6598317e9b9a7d0cfcd.

2018-03-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
index 067b5c4..beb9075 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html
@@ -126,7 +126,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class RegionCoprocessorHost.RegionEnvironment
+private static class RegionCoprocessorHost.RegionEnvironment
 extends BaseEnvironmentRegionCoprocessor
 implements RegionCoprocessorEnvironment
 Encapsulation of the environment of each coprocessor
@@ -303,7 +303,7 @@ implements 
 
 region
-privateRegion region
+privateRegion region
 
 
 
@@ -312,7 +312,7 @@ implements 
 
 sharedData
-https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
 title="class or interface in java.util.concurrent">ConcurrentMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object sharedData
+https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
 title="class or interface in java.util.concurrent">ConcurrentMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object sharedData
 
 
 
@@ -321,7 +321,7 @@ implements 
 
 metricRegistry
-private finalMetricRegistry metricRegistry
+private finalMetricRegistry metricRegistry
 
 
 
@@ -330,7 +330,7 @@ implements 
 
 services
-private finalRegionServerServices services
+private finalRegionServerServices services
 
 
 
@@ -347,7 +347,7 @@ implements 
 
 RegionEnvironment
-publicRegionEnvironment(RegionCoprocessorimpl,
+publicRegionEnvironment(RegionCoprocessorimpl,
  intpriority,
  intseq,
  org.apache.hadoop.conf.Configurationconf,
@@ -376,7 +376,7 @@ implements 
 
 getRegion
-publicRegiongetRegion()
+publicRegiongetRegion()
 
 Specified by:
 getRegionin
 interfaceRegionCoprocessorEnvironment
@@ -391,7 +391,7 @@ implements 
 
 getOnlineRegions
-publicOnlineRegionsgetOnlineRegions()
+publicOnlineRegionsgetOnlineRegions()
 
 Specified by:
 getOnlineRegionsin
 interfaceRegionCoprocessorEnvironment
@@ -406,7 +406,7 @@ implements 
 
 getConnection
-publicConnectiongetConnection()
+publicConnectiongetConnection()
 Description copied from 
interface:RegionCoprocessorEnvironment
 Returns the hosts' Connection to the Cluster. Do not 
close! This is a shared connection
  with the hosting server. Throws https://docs.oracle.com/javase/8/docs/api/java/lang/UnsupportedOperationException.html?is-external=true;
 title="class or interface in 
java.lang">UnsupportedOperationException if you try to close
@@ -445,7 +445,7 @@ implements 
 
 createConnection
-publicConnectioncreateConnection(org.apache.hadoop.conf.Configurationconf)
+publicConnectioncreateConnection(org.apache.hadoop.conf.Configurationconf)
 throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Description copied from 
interface:RegionCoprocessorEnvironment
 Creates a cluster connection using the passed Configuration.
@@ -481,7 +481,7 @@ implements 
 
 getServerName
-publicServerNamegetServerName()
+publicServerNamegetServerName()
 
 Specified by:
 getServerNamein
 interfaceRegionCoprocessorEnvironment
@@ -496,7 +496,7 @@ implements 
 
 shutdown
-publicvoidshutdown()
+publicvoidshutdown()
 Description copied from 
class:BaseEnvironment
 Clean up the environment
 
@@ -511,7 +511,7 @@ implements 
 
 getSharedData
-publichttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
 title="class or interface in java.util.concurrent">ConcurrentMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">ObjectgetSharedData()
+publichttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
 title="class or interface in 

[27/51] [partial] hbase-site git commit: Published site at 31da4d0bce69b3a47066a5df675756087ce4dc60.

2018-03-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
index fa987ee..a13341f 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html
@@ -317,7 +317,7 @@ extends AbstractStateMachineTableProcedure
-acquireLock,
 checkTableModifiable,
 getRegionDir,
 getUser,
 releaseLock,
 <
 a 
href="../../../../../../org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html#releaseSyncLatch--">releaseSyncLatch,
 setUser
+acquireLock,
 checkTableModifiable,
 getRegionDir,
 getUser,
 preflightChecks, releaseLock,
 releaseSyncLatch,
 setUser
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html
index 72ff612..671b283 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html
@@ -305,7 +305,7 @@ extends AbstractStateMachineTableProcedure
-getRegionDir,
 getUser,
 releaseSyncLatch,
 setUser
+getRegionDir,
 getUser,
 preflightChecks,
 releaseSyncLatch,
 setUser
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
index 347cea7..2881954 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
@@ -129,11 +129,13 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class MergeTableRegionsProcedure
+public class MergeTableRegionsProcedure
 extends AbstractStateMachineTableProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsState
 The procedure to Merge a region in a table.
  This procedure takes an exclusive table lock since it is working over 
multiple regions.
- It holds the lock for the life of the procedure.
+ It holds the lock for the life of the procedure.
+ Throws exception on construction if determines context hostile to merge 
(cluster going
+ down or master is shutting down or table is disabled).
 
 
 
@@ -520,7 +522,7 @@ extends AbstractStateMachineTableProcedure
-checkTableModifiable,
 getRegionDir,
 getUser,
 releaseSyncLatch,
 setUser
+checkTableModifiable,
 getRegionDir,
 getUser,
 preflightChecks,
 releaseSyncLatch,
 setUser
 
 
 
@@ -563,7 +565,7 @@ extends 
 
 LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
 
 
 
@@ -572,7 +574,7 @@ extends 
 
 traceEnabled
-privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
+privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
 
 
 
@@ -581,7 +583,7 @@ extends 
 
 lock
-private volatileboolean lock
+private volatileboolean lock
 
 
 
@@ -590,7 +592,7 @@ extends 
 
 regionLocation
-privateServerName regionLocation
+privateServerName regionLocation
 
 
 
@@ -599,7 +601,7 @@ extends 
 
 regionsToMerge
-privateRegionInfo[] regionsToMerge
+privateRegionInfo[] regionsToMerge
 
 
 
@@ -608,7 +610,7 @@ extends 
 
 mergedRegion
-privateRegionInfo mergedRegion
+privateRegionInfo mergedRegion
 
 
 
@@ -617,7 +619,7 @@ extends 
 
 forcible
-privateboolean forcible
+privateboolean forcible
 
 
 
@@ -634,7 +636,7 @@ extends 
 
 MergeTableRegionsProcedure
-publicMergeTableRegionsProcedure()
+publicMergeTableRegionsProcedure()
 
 
 
@@ -643,7 +645,7 @@ extends 
 
 MergeTableRegionsProcedure
-publicMergeTableRegionsProcedure(MasterProcedureEnvenv,
+publicMergeTableRegionsProcedure(MasterProcedureEnvenv,
   RegionInforegionToMergeA,
   RegionInforegionToMergeB)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in 

[27/51] [partial] hbase-site git commit: Published site at 6b77786dfc46d25ac5bb5f1c8a4a9eb47b52a604.

2018-03-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
index 71b940f..7ac76dd 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html
@@ -106,11 +106,11 @@
 
 
 private AsyncConnectionImpl
-RawAsyncTableImpl.conn
+AsyncClientScanner.conn
 
 
 private AsyncConnectionImpl
-AsyncBatchRpcRetryingCaller.conn
+AsyncRpcRetryingCallerFactory.conn
 
 
 private AsyncConnectionImpl
@@ -118,19 +118,19 @@
 
 
 private AsyncConnectionImpl
-RegionCoprocessorRpcChannelImpl.conn
+RawAsyncTableImpl.conn
 
 
-protected AsyncConnectionImpl
-AsyncRpcRetryingCaller.conn
+private AsyncConnectionImpl
+RegionCoprocessorRpcChannelImpl.conn
 
 
 private AsyncConnectionImpl
-AsyncClientScanner.conn
+AsyncBatchRpcRetryingCaller.conn
 
 
-private AsyncConnectionImpl
-AsyncRpcRetryingCallerFactory.conn
+protected AsyncConnectionImpl
+AsyncRpcRetryingCaller.conn
 
 
 private AsyncConnectionImpl

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
index e71ca45..d6b1759 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html
@@ -105,13 +105,13 @@
 
 
 
-private AsyncMasterRequestRpcRetryingCaller.CallableT
-AsyncMasterRequestRpcRetryingCaller.callable
-
-
 private AsyncMasterRequestRpcRetryingCaller.CallableT
 AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder.callable
 
+
+private AsyncMasterRequestRpcRetryingCaller.CallableT
+AsyncMasterRequestRpcRetryingCaller.callable
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
index 5f86673..f2374f5 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html
@@ -106,11 +106,11 @@
 
 
 private AsyncProcess
-BufferedMutatorImpl.ap
+HTableMultiplexer.FlushWorker.ap
 
 
 private AsyncProcess
-HTableMultiplexer.FlushWorker.ap
+BufferedMutatorImpl.ap
 
 
 private AsyncProcess
@@ -137,11 +137,11 @@
 
 
 AsyncProcess
-ClusterConnection.getAsyncProcess()
+ConnectionImplementation.getAsyncProcess()
 
 
 AsyncProcess
-ConnectionImplementation.getAsyncProcess()
+ClusterConnection.getAsyncProcess()
 
 
 (package private) AsyncProcess

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html
index c610e19..9a8d746 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html
@@ -106,11 +106,11 @@
 
 
 private AsyncRegionLocator
-AsyncConnectionImpl.locator
+AsyncTableRegionLocatorImpl.locator
 
 
 private AsyncRegionLocator
-AsyncTableRegionLocatorImpl.locator
+AsyncConnectionImpl.locator
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html
index 447fe97..e4ecf34 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html
@@ -126,13 +126,13 @@
 
 
 
-(package private) AsyncRegistry
-AsyncConnectionImpl.registry
-
-
 private AsyncRegistry
 AsyncMetaRegionLocator.registry
 
+
+(package private) AsyncRegistry
+AsyncConnectionImpl.registry
+
 
 
 


[27/51] [partial] hbase-site git commit: Published site at 1384da71375427b522b09f06862bb5d629cef52f.

2018-03-06 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
index d788050..bba78a1 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
@@ -2052,119 +2052,119 @@ service.
 
 
 private TableName
-RegionCoprocessorRpcChannel.table
+SnapshotDescription.table
 
 
 private TableName
-SnapshotDescription.table
+RegionCoprocessorRpcChannel.table
 
 
 private TableName
-HRegionLocator.tableName
+RawAsyncTableImpl.tableName
 
 
 private TableName
-ScannerCallableWithReplicas.tableName
+RegionServerCallable.tableName
 
 
 protected TableName
-ClientScanner.tableName
+RegionAdminServiceCallable.tableName
 
 
 private TableName
-AsyncClientScanner.tableName
+BufferedMutatorImpl.tableName
 
 
 private TableName
-AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.tableName
+AsyncProcessTask.tableName
 
 
 private TableName
-AsyncRpcRetryingCallerFactory.BatchCallerBuilder.tableName
+AsyncProcessTask.Builder.tableName
 
 
 private TableName
-RegionInfoBuilder.tableName
+AsyncRequestFutureImpl.tableName
 
 
-private TableName
-RegionInfoBuilder.MutableRegionInfo.tableName
+protected TableName
+TableBuilderBase.tableName
 
 
 private TableName
-RawAsyncTableImpl.tableName
+AsyncBatchRpcRetryingCaller.tableName
 
 
 private TableName
-RegionCoprocessorRpcChannelImpl.tableName
+RegionInfoBuilder.tableName
 
 
 private TableName
-AsyncTableRegionLocatorImpl.tableName
+RegionInfoBuilder.MutableRegionInfo.tableName
 
 
-protected TableName
-RegionAdminServiceCallable.tableName
+private TableName
+HTable.tableName
 
 
 private TableName
-HTable.tableName
+TableState.tableName
 
 
-private TableName
-BufferedMutatorImpl.tableName
+protected TableName
+RpcRetryingCallerWithReadReplicas.tableName
 
 
-private TableName
-AsyncBatchRpcRetryingCaller.tableName
+protected TableName
+AsyncTableBuilderBase.tableName
 
 
 private TableName
-BufferedMutatorParams.tableName
+AsyncSingleRequestRpcRetryingCaller.tableName
 
 
 private TableName
-HBaseAdmin.TableFuture.tableName
+ScannerCallableWithReplicas.tableName
 
 
-private TableName
-AsyncRequestFutureImpl.tableName
+protected TableName
+RawAsyncHBaseAdmin.TableProcedureBiConsumer.tableName
 
 
 private TableName
-AsyncProcessTask.tableName
+AsyncTableRegionLocatorImpl.tableName
 
 
 private TableName
-AsyncProcessTask.Builder.tableName
+HBaseAdmin.TableFuture.tableName
 
 
-protected TableName
-RawAsyncHBaseAdmin.TableProcedureBiConsumer.tableName
+private TableName
+RegionCoprocessorRpcChannelImpl.tableName
 
 
-private TableName
-RegionServerCallable.tableName
+protected TableName
+ClientScanner.tableName
 
 
 private TableName
-AsyncSingleRequestRpcRetryingCaller.tableName
+BufferedMutatorParams.tableName
 
 
-protected TableName
-TableBuilderBase.tableName
+private TableName
+AsyncClientScanner.tableName
 
 
-protected TableName
-RpcRetryingCallerWithReadReplicas.tableName
+private TableName
+AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.tableName
 
 
-protected TableName
-AsyncTableBuilderBase.tableName
+private TableName
+AsyncRpcRetryingCallerFactory.BatchCallerBuilder.tableName
 
 
 private TableName
-TableState.tableName
+HRegionLocator.tableName
 
 
 
@@ -2206,83 +2206,83 @@ service.
 
 
 TableName
-AsyncTable.getName()
-Gets the fully qualified table name instance of this 
table.
-
+RawAsyncTableImpl.getName()
 
 
 TableName
-Table.getName()
+RegionLocator.getName()
 Gets the fully qualified table name instance of this 
table.
 
 
 
 TableName
-HRegionLocator.getName()
+BufferedMutatorImpl.getName()
 
 
 TableName
-AsyncTableRegionLocator.getName()
-Gets the fully qualified table name instance of the table 
whose region we want to locate.
+BufferedMutator.getName()
+Gets the fully qualified table name instance of the table 
that this BufferedMutator writes to.
 
 
 
 TableName
-AsyncTableImpl.getName()
+HTable.getName()
 
 
 TableName
-RawAsyncTableImpl.getName()
+AsyncBufferedMutator.getName()
+Gets the fully qualified table name instance of the table 
that this
+ AsyncBufferedMutator writes to.
+
 
 
 TableName
-AsyncTableRegionLocatorImpl.getName()
+Table.getName()
+Gets the fully qualified table name instance of this 
table.
+
 
 
 TableName
-BufferedMutator.getName()
-Gets the fully qualified table name instance of the table 
that this BufferedMutator writes to.
-
+AsyncTableImpl.getName()
 
 
 TableName
-RegionLocator.getName()
+AsyncTable.getName()
 Gets the fully qualified table name instance of this 
table.
 
 
 
 TableName
-AsyncBufferedMutatorImpl.getName()
+AsyncTableRegionLocatorImpl.getName()
 
 
 TableName
-HTable.getName()
+AsyncTableRegionLocator.getName()
+Gets the fully qualified table 

[27/51] [partial] hbase-site git commit: Published site at b7b86839250bf9b295ebc1948826f43a88736d6c.

2018-03-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
 
b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
index 2b5efbb..f0821d4 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html
@@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
 
 
 PrevClass
-NextClass
+NextClass
 
 
 Frames
@@ -74,7 +74,7 @@ var activeTableTab = "activeTableTab";
 
 
 Summary:
-Nested|
+Nested|
 Field|
 Constr|
 Method
@@ -126,7 +126,8 @@ var activeTableTab = "activeTableTab";
  @InterfaceStability.Evolving
 public class ReplicationSourceWALReader
 extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true;
 title="class or interface in java.lang">Thread
-Reads and filters WAL entries, groups the filtered entries 
into batches, and puts the batches onto a queue
+Reads and filters WAL entries, groups the filtered entries 
into batches, and puts the batches
+ onto a queue
 
 
 
@@ -139,19 +140,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 Nested Class Summary
-
-Nested Classes
-
-Modifier and Type
-Class and Description
-
-
-(package private) static class
-ReplicationSourceWALReader.WALEntryBatch
-Holds a batch of WAL entries to replicate, along with some 
statistics
-
-
-
 
 
 
@@ -182,7 +170,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 currentPosition
 
 
-protected https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true;
 title="class or interface in java.util.concurrent">BlockingQueueReplicationSourceWALReader.WALEntryBatch
+protected https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true;
 title="class or interface in java.util.concurrent">BlockingQueueWALEntryBatch
 entryBatchQueue
 
 
@@ -222,18 +210,22 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 replicationBatchSizeCapacity
 
 
+private SerialReplicationChecker
+serialReplicationChecker
+
+
 private long
 sleepForRetries
 
-
+
 private ReplicationSource
 source
 
-
+
 private long
 totalBufferQuota
 
-
+
 private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true;
 title="class or interface in 
java.util.concurrent.atomic">AtomicLong
 totalBufferUsed
 
@@ -317,8 +309,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 protected void
-handleEmptyWALEntryBatch(ReplicationSourceWALReader.WALEntryBatchbatch,
-
org.apache.hadoop.fs.PathcurrentPath)
+handleEmptyWALEntryBatch(org.apache.hadoop.fs.PathcurrentPath)
 
 
 private void
@@ -329,7 +320,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 isReaderRunning()
 
 
-private ReplicationSourceWALReader.WALEntryBatch
+private WALEntryBatch
 readWALEntries(WALEntryStreamentryStream)
 
 
@@ -341,7 +332,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 setReaderRunning(booleanreaderRunning)
 
 
-ReplicationSourceWALReader.WALEntryBatch
+WALEntryBatch
 take()
 Retrieves the next batch of WAL entries from the queue, 
waiting up to the specified time for a
  batch to become available
@@ -349,9 +340,8 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 private void
-updateBatchStats(ReplicationSourceWALReader.WALEntryBatchbatch,
+updateBatchStats(WALEntryBatchbatch,
 WAL.Entryentry,
-longentryPosition,
 longentrySize)
 
 
@@ -443,7 +433,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 entryBatchQueue
-protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true;
 title="class or interface in java.util.concurrent">BlockingQueueReplicationSourceWALReader.WALEntryBatch
 entryBatchQueue
+protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true;
 title="class or interface in java.util.concurrent">BlockingQueueWALEntryBatch entryBatchQueue
 
 
 
@@ -521,12 +511,21 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 
-
+
 
 totalBufferQuota
 privatelong totalBufferQuota
 
 
+
+
+
+
+
+serialReplicationChecker
+private finalSerialReplicationChecker 
serialReplicationChecker
+
+
 
 
 
@@ -541,7 +540,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html
 
 
 ReplicationSourceWALReader

[27/51] [partial] hbase-site git commit: Published site at 1d25b60831b8cc8f7ad5fd366f1867de5c20d2f3.

2018-03-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html 
b/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html
index 6ea4d25..1900827 100644
--- a/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html
+++ b/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html
@@ -117,7 +117,7 @@
 
 
 
-static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
+static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
 ServerName.EMPTY_SERVER_LIST
 
 
@@ -166,7 +166,7 @@
 
 
 static ServerName
-ServerName.parseServerName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringstr)
+ServerName.parseServerName(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringstr)
 
 
 static ServerName
@@ -177,13 +177,13 @@
 
 
 static ServerName
-ServerName.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringserverName)
+ServerName.valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringserverName)
 Retrieve an instance of ServerName.
 
 
 
 static ServerName
-ServerName.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringhostname,
+ServerName.valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringhostname,
intport,
longstartcode)
 Retrieve an instance of ServerName.
@@ -191,7 +191,7 @@
 
 
 static ServerName
-ServerName.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringhostAndPort,
+ServerName.valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringhostAndPort,
longstartCode)
 Retrieve an instance of ServerName.
 
@@ -206,17 +206,17 @@
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
+https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
 ClusterMetrics.getBackupMasterNames()
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
+https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
 ClusterStatus.getBackupMasterNames()
 Deprecated.
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
+https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
 ClusterStatus.getBackupMasters()
 Deprecated.
 As of release 2.0.0, this 
will be removed in HBase 3.0.0
@@ -225,27 +225,27 @@
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
+https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
 ClusterMetrics.getDeadServerNames()
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
+https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListServerName
 ClusterStatus.getDeadServerNames()
 Deprecated.
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">MapServerName,ServerMetrics
+https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">MapServerName,ServerMetrics
 ClusterMetrics.getLiveServerMetrics()
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">MapServerName,ServerMetrics
+https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">MapServerName,ServerMetrics
 ClusterStatus.getLiveServerMetrics()
 Deprecated.
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface 

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html 
b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
index b419e6d..8e88e74 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
@@ -2585,7 +2585,7 @@ implements 
 
 stopProcedureExecutor
-privatevoidstopProcedureExecutor()
+privatevoidstopProcedureExecutor()
 
 
 
@@ -2594,7 +2594,7 @@ implements 
 
 stopChores
-privatevoidstopChores()
+privatevoidstopChores()
 
 
 
@@ -2603,7 +2603,7 @@ implements 
 
 getRemoteInetAddress
-http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true;
 title="class or interface in java.net">InetAddressgetRemoteInetAddress(intport,
+http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true;
 title="class or interface in java.net">InetAddressgetRemoteInetAddress(intport,
  longserverStartCode)
   throws http://docs.oracle.com/javase/8/docs/api/java/net/UnknownHostException.html?is-external=true;
 title="class or interface in java.net">UnknownHostException
 
@@ -2620,7 +2620,7 @@ implements 
 
 getMaxBalancingTime
-privateintgetMaxBalancingTime()
+privateintgetMaxBalancingTime()
 
 Returns:
 Maximum time we should run balancer for
@@ -2633,7 +2633,7 @@ implements 
 
 getMaxRegionsInTransition
-privateintgetMaxRegionsInTransition()
+privateintgetMaxRegionsInTransition()
 
 Returns:
 Maximum number of regions in transition
@@ -2646,7 +2646,7 @@ implements 
 
 balanceThrottling
-privatevoidbalanceThrottling(longnextBalanceStartTime,
+privatevoidbalanceThrottling(longnextBalanceStartTime,
intmaxRegionsInTransition,
longcutoffTime)
 It first sleep to the next balance plan start time. 
Meanwhile, throttling by the max
@@ -2665,7 +2665,7 @@ implements 
 
 balance
-publicbooleanbalance()
+publicbooleanbalance()
 throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
 Throws:
@@ -2679,7 +2679,7 @@ implements 
 
 balance
-publicbooleanbalance(booleanforce)
+publicbooleanbalance(booleanforce)
 throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
 Throws:
@@ -2693,7 +2693,7 @@ implements 
 
 getRegionNormalizer
-publicRegionNormalizergetRegionNormalizer()
+publicRegionNormalizergetRegionNormalizer()
 
 Specified by:
 getRegionNormalizerin
 interfaceMasterServices
@@ -2708,7 +2708,7 @@ implements 
 
 normalizeRegions
-publicbooleannormalizeRegions()
+publicbooleannormalizeRegions()
  throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Perform normalization of cluster (invoked by RegionNormalizerChore).
 
@@ -2727,7 +2727,7 @@ implements 
 
 getClientIdAuditPrefix
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetClientIdAuditPrefix()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetClientIdAuditPrefix()
 
 Specified by:
 getClientIdAuditPrefixin
 interfaceMasterServices
@@ -2742,7 +2742,7 @@ implements 
 
 setCatalogJanitorEnabled
-publicvoidsetCatalogJanitorEnabled(booleanb)
+publicvoidsetCatalogJanitorEnabled(booleanb)
 Switch for the background CatalogJanitor thread.
  Used for testing.  The thread will continue to run.  It will just be a noop
  if disabled.
@@ -2758,7 +2758,7 @@ implements 
 
 mergeRegions
-publiclongmergeRegions(RegionInfo[]regionsToMerge,
+publiclongmergeRegions(RegionInfo[]regionsToMerge,
  booleanforcible,
  longnonceGroup,
  longnonce)
@@ -2786,7 +2786,7 @@ implements 
 
 splitRegion
-publiclongsplitRegion(RegionInforegionInfo,
+publiclongsplitRegion(RegionInforegionInfo,
 byte[]splitRow,
 longnonceGroup,
 longnonce)
@@ -2814,7 +2814,7 @@ implements 
 
 move
-publicvoidmove(byte[]encodedRegionName,
+publicvoidmove(byte[]encodedRegionName,
  byte[]destServerName)
   throws HBaseIOException
 
@@ -2829,7 +2829,7 @@ implements 
 
 createTable
-publiclongcreateTable(TableDescriptortableDescriptor,
+publiclongcreateTable(TableDescriptortableDescriptor,
 byte[][]splitKeys,
 longnonceGroup,
 longnonce)
@@ -2854,7 +2854,7 @@ implements 
 
 

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
index d143ef8..4583895 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html
@@ -258,7 +258,7 @@
 250
 251@Override
 252public long heapSize() {
-253  long sum = HEAP_SIZE_OVERHEAD + 
estimatedHeapSizeOf(cell);
+253  long sum = HEAP_SIZE_OVERHEAD + 
estimatedSizeOfCell(cell);
 254  if (this.tags != null) {
 255sum += 
ClassSize.sizeOf(this.tags);
 256  }
@@ -454,7 +454,7 @@
 446
 447@Override
 448public long heapSize() {
-449  long sum = HEAP_SIZE_OVERHEAD + 
estimatedHeapSizeOf(cell);
+449  long sum = HEAP_SIZE_OVERHEAD + 
estimatedSizeOfCell(cell);
 450  // this.tags is on heap byte[]
 451  if (this.tags != null) {
 452sum += 
ClassSize.sizeOf(this.tags);
@@ -2791,192 +2791,193 @@
 2783   * {@link HeapSize} we call {@link 
HeapSize#heapSize()} so cell can give a correct value. In other
 2784   * cases we just consider the bytes 
occupied by the cell components ie. row, CF, qualifier,
 2785   * timestamp, type, value and tags.
-2786   * @param cell
-2787   * @return estimate of the heap 
space
-2788   */
-2789  public static long 
estimatedHeapSizeOf(final Cell cell) {
-2790if (cell instanceof HeapSize) {
-2791  return ((HeapSize) 
cell).heapSize();
-2792}
-2793// TODO: Add sizing of references 
that hold the row, family, etc., arrays.
-2794return 
estimatedSerializedSizeOf(cell);
-2795  }
-2796
-2797  /**
-2798   * This method exists just to 
encapsulate how we serialize keys. To be replaced by a factory that
-2799   * we query to figure what the Cell 
implementation is and then, what serialization engine to use
-2800   * and further, how to serialize the 
key for inclusion in hfile index. TODO.
-2801   * @param cell
-2802   * @return The key portion of the Cell 
serialized in the old-school KeyValue way or null if passed
-2803   * a null 
codecell/code
-2804   */
-2805  public static byte[] 
getCellKeySerializedAsKeyValueKey(final Cell cell) {
-2806if (cell == null) return null;
-2807byte[] b = new 
byte[KeyValueUtil.keyLength(cell)];
-2808KeyValueUtil.appendKeyTo(cell, b, 
0);
-2809return b;
-2810  }
-2811
-2812  /**
-2813   * Create a Cell that is smaller than 
all other possible Cells for the given Cell's row.
-2814   * @param cell
-2815   * @return First possible Cell on 
passed Cell's row.
-2816   */
-2817  public static Cell 
createFirstOnRow(final Cell cell) {
-2818if (cell instanceof 
ByteBufferExtendedCell) {
-2819  return new 
FirstOnRowByteBufferExtendedCell(
-2820  ((ByteBufferExtendedCell) 
cell).getRowByteBuffer(),
-2821  ((ByteBufferExtendedCell) 
cell).getRowPosition(), cell.getRowLength());
-2822}
-2823return new 
FirstOnRowCell(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength());
-2824  }
-2825
-2826  public static Cell 
createFirstOnRow(final byte[] row, int roffset, short rlength) {
-2827return new FirstOnRowCell(row, 
roffset, rlength);
-2828  }
-2829
-2830  public static Cell 
createFirstOnRow(final byte[] row, final byte[] family, final byte[] col) {
-2831return createFirstOnRow(row, 0, 
(short) row.length, family, 0, (byte) family.length, col, 0,
-2832col.length);
-2833  }
-2834
-2835  public static Cell 
createFirstOnRow(final byte[] row, int roffset, short rlength,
-2836  final byte[] family, int foffset, 
byte flength, final byte[] col, int coffset, int clength) {
-2837return new FirstOnRowColCell(row, 
roffset, rlength, family, foffset, flength, col, coffset,
-2838clength);
-2839  }
-2840
-2841  public static Cell 
createFirstOnRow(final byte[] row) {
-2842return createFirstOnRow(row, 0, 
(short) row.length);
-2843  }
-2844
-2845  public static Cell 
createFirstOnRowFamily(Cell cell, byte[] fArray, int foff, int flen) {
-2846if (cell instanceof 
ByteBufferExtendedCell) {
-2847  return new 
FirstOnRowColByteBufferExtendedCell(
-2848  ((ByteBufferExtendedCell) 
cell).getRowByteBuffer(),
-2849  ((ByteBufferExtendedCell) 
cell).getRowPosition(), cell.getRowLength(),
-2850  ByteBuffer.wrap(fArray), foff, 
(byte) flen, HConstants.EMPTY_BYTE_BUFFER, 0, 0);
-2851}
-2852return new 
FirstOnRowColCell(cell.getRowArray(), cell.getRowOffset(), 
cell.getRowLength(),
-2853fArray, foff, (byte) flen, 
HConstants.EMPTY_BYTE_ARRAY, 0, 0);
-2854  }
-2855
-2856  public static Cell 
createFirstOnRowCol(final Cell cell) {
-2857if (cell 

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-17 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
index cda7c78..93f650f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html
@@ -269,7 +269,7 @@
 261this.rpcControllerFactory = 
RpcControllerFactory.instantiate(conf);
 262this.rpcCallerFactory = 
RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats);
 263this.backoffPolicy = 
ClientBackoffPolicyFactory.create(conf);
-264this.asyncProcess = new 
AsyncProcess(this, conf, rpcCallerFactory, false, rpcControllerFactory);
+264this.asyncProcess = new 
AsyncProcess(this, conf, rpcCallerFactory, rpcControllerFactory);
 265if 
(conf.getBoolean(CLIENT_SIDE_METRICS_ENABLED_KEY, false)) {
 266  this.metrics = new 
MetricsConnection(this);
 267} else {

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
index cda7c78..93f650f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html
@@ -269,7 +269,7 @@
 261this.rpcControllerFactory = 
RpcControllerFactory.instantiate(conf);
 262this.rpcCallerFactory = 
RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats);
 263this.backoffPolicy = 
ClientBackoffPolicyFactory.create(conf);
-264this.asyncProcess = new 
AsyncProcess(this, conf, rpcCallerFactory, false, rpcControllerFactory);
+264this.asyncProcess = new 
AsyncProcess(this, conf, rpcCallerFactory, rpcControllerFactory);
 265if 
(conf.getBoolean(CLIENT_SIDE_METRICS_ENABLED_KEY, false)) {
 266  this.metrics = new 
MetricsConnection(this);
 267} else {

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
index cda7c78..93f650f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
@@ -269,7 +269,7 @@
 261this.rpcControllerFactory = 
RpcControllerFactory.instantiate(conf);
 262this.rpcCallerFactory = 
RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats);
 263this.backoffPolicy = 
ClientBackoffPolicyFactory.create(conf);
-264this.asyncProcess = new 
AsyncProcess(this, conf, rpcCallerFactory, false, rpcControllerFactory);
+264this.asyncProcess = new 
AsyncProcess(this, conf, rpcCallerFactory, rpcControllerFactory);
 265if 
(conf.getBoolean(CLIENT_SIDE_METRICS_ENABLED_KEY, false)) {
 266  this.metrics = new 
MetricsConnection(this);
 267} else {

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
index cda7c78..93f650f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html
@@ -269,7 +269,7 @@
 261this.rpcControllerFactory = 
RpcControllerFactory.instantiate(conf);
 262this.rpcCallerFactory = 
RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats);
 263this.backoffPolicy = 

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
index 90f52b0..2ac1b78 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html
@@ -106,7 +106,7 @@
 
 
 private RegionLocateType
-AsyncSingleRequestRpcRetryingCaller.locateType
+AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.locateType
 
 
 RegionLocateType
@@ -114,7 +114,7 @@
 
 
 private RegionLocateType
-AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.locateType
+AsyncSingleRequestRpcRetryingCaller.locateType
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
index e062eb5..fbe0658 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
@@ -230,14 +230,14 @@ service.
 
 
 private RegionLocator
-HFileOutputFormat2.TableInfo.regionLocator
-
-
-private RegionLocator
 TableInputFormatBase.regionLocator
 The RegionLocator of the 
table.
 
 
+
+private RegionLocator
+HFileOutputFormat2.TableInfo.regionLocator
+
 
 
 
@@ -248,15 +248,15 @@ service.
 
 
 
-RegionLocator
-HFileOutputFormat2.TableInfo.getRegionLocator()
-
-
 protected RegionLocator
 TableInputFormatBase.getRegionLocator()
 Allows subclasses to get the RegionLocator.
 
 
+
+RegionLocator
+HFileOutputFormat2.TableInfo.getRegionLocator()
+
 
 
 



[27/51] [partial] hbase-site git commit: Published site at .

2018-02-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
index ded834e..9c4acf5 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html
@@ -330,8 +330,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncHBaseAdmin.addColumnFamily(TableNametableName,
-   ColumnFamilyDescriptorcolumnFamily)
+AsyncAdmin.addColumnFamily(TableNametableName,
+   ColumnFamilyDescriptorcolumnFamily)
+Add a column family to an existing table.
+
 
 
 void
@@ -341,20 +343,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-void
-HBaseAdmin.addColumnFamily(TableNametableName,
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
+RawAsyncHBaseAdmin.addColumnFamily(TableNametableName,
ColumnFamilyDescriptorcolumnFamily)
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncAdmin.addColumnFamily(TableNametableName,
-   ColumnFamilyDescriptorcolumnFamily)
-Add a column family to an existing table.
-
+void
+HBaseAdmin.addColumnFamily(TableNametableName,
+   ColumnFamilyDescriptorcolumnFamily)
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-RawAsyncHBaseAdmin.addColumnFamily(TableNametableName,
+AsyncHBaseAdmin.addColumnFamily(TableNametableName,
ColumnFamilyDescriptorcolumnFamily)
 
 
@@ -396,8 +396,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncHBaseAdmin.modifyColumnFamily(TableNametableName,
-  ColumnFamilyDescriptorcolumnFamily)
+AsyncAdmin.modifyColumnFamily(TableNametableName,
+  ColumnFamilyDescriptorcolumnFamily)
+Modify an existing column family on a table.
+
 
 
 void
@@ -407,20 +409,18 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-void
-HBaseAdmin.modifyColumnFamily(TableNametableName,
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
+RawAsyncHBaseAdmin.modifyColumnFamily(TableNametableName,
   ColumnFamilyDescriptorcolumnFamily)
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncAdmin.modifyColumnFamily(TableNametableName,
-  ColumnFamilyDescriptorcolumnFamily)
-Modify an existing column family on a table.
-
+void
+HBaseAdmin.modifyColumnFamily(TableNametableName,
+  ColumnFamilyDescriptorcolumnFamily)
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
index 53d69df..597d1ba 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
@@ -355,24 +355,24 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 void
-BufferedMutatorImpl.mutate(Mutationm)
-
-
-void
 BufferedMutator.mutate(Mutationmutation)
 Sends a Mutation to 
the table.
 
 
+
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
+AsyncBufferedMutatorImpl.mutate(Mutationmutation)
+
 
+void
+BufferedMutatorImpl.mutate(Mutationm)
+
+
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
 AsyncBufferedMutator.mutate(Mutationmutation)
 Sends a Mutation to 
the table.
 
 
-
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncBufferedMutatorImpl.mutate(Mutationmutation)
-
 
 
 
@@ -390,24 +390,24 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 void
-BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationms)
-
-
-void
 BufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
 Send some Mutations to 
the table.
 
 
+
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
+AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
+
 
+void
+BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationms)
+
+
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
 AsyncBufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
 Send some Mutations to 
the table.
 
 
-
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
-
 
 static RowMutations
 RowMutations.of(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
@@ -543,15 +543,15 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 MutationSerialization.getDeserializer(http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">ClassMutationc)
 
 

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-13 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
index 1d50582..7eb7661 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html
@@ -104,14 +104,14 @@
 
 
 void
-HMaster.checkTableModifiable(TableNametableName)
-
-
-void
 MasterServices.checkTableModifiable(TableNametableName)
 Check table is modifiable; i.e.
 
 
+
+void
+HMaster.checkTableModifiable(TableNametableName)
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
index 5eebc7c..b1a475b 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html
@@ -144,6 +144,19 @@
 
 
 Uses of TableNotFoundException in org.apache.hadoop.hbase.master
+
+Subclasses of TableNotFoundException in org.apache.hadoop.hbase.master
+
+Modifier and Type
+Class and Description
+
+
+
+static class
+TableStateManager.TableStateNotFoundException
+
+
+
 
 Methods in org.apache.hadoop.hbase.master
 that throw TableNotFoundException
 
@@ -157,14 +170,14 @@
 
 
 void
-HMaster.checkTableModifiable(TableNametableName)
-
-
-void
 MasterServices.checkTableModifiable(TableNametableName)
 Check table is modifiable; i.e.
 
 
+
+void
+HMaster.checkTableModifiable(TableNametableName)
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
index 594e74b..90f6967 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html
@@ -243,6 +243,10 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTag
+PrivateCellUtil.getTags(Cellcell)
+
+
+static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTag
 CellUtil.getTags(Cellcell)
 Deprecated.
 As of release 2.0.0, this 
will be removed in HBase 3.0.0.
@@ -250,10 +254,6 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-
-static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTag
-PrivateCellUtil.getTags(Cellcell)
-
 
 static http://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true;
 title="class or interface in java.util">IteratorTag
 CellUtil.tagsIterator(byte[]tags,
@@ -395,6 +395,11 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 static Cell
+PrivateCellUtil.createCell(Cellcell,
+  http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTagtags)
+
+
+static Cell
 CellUtil.createCell(Cellcell,
   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTagtags)
 Deprecated.
@@ -402,11 +407,6 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-
-static Cell
-PrivateCellUtil.createCell(Cellcell,
-  http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTagtags)
-
 
 static byte[]
 TagUtil.fromList(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTagtags)
@@ -415,16 +415,16 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-RawCellBuilder
-RawCellBuilder.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTagtags)
+ExtendedCellBuilder
+ExtendedCellBuilderImpl.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListTagtags)
 
 
-ExtendedCellBuilder

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
index 53d69df..597d1ba 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html
@@ -355,24 +355,24 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 void
-BufferedMutatorImpl.mutate(Mutationm)
-
-
-void
 BufferedMutator.mutate(Mutationmutation)
 Sends a Mutation to 
the table.
 
 
+
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
+AsyncBufferedMutatorImpl.mutate(Mutationmutation)
+
 
+void
+BufferedMutatorImpl.mutate(Mutationm)
+
+
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
 AsyncBufferedMutator.mutate(Mutationmutation)
 Sends a Mutation to 
the table.
 
 
-
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncBufferedMutatorImpl.mutate(Mutationmutation)
-
 
 
 
@@ -390,24 +390,24 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 void
-BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationms)
-
-
-void
 BufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
 Send some Mutations to 
the table.
 
 
+
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
+AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
+
 
+void
+BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationms)
+
+
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
 AsyncBufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
 Send some Mutations to 
the table.
 
 
-
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
-AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
-
 
 static RowMutations
 RowMutations.of(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Mutationmutations)
@@ -543,15 +543,15 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 MutationSerialization.getDeserializer(http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
 title="class or interface in java.lang">ClassMutationc)
 
 

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
index 232a8b4..aeaf9fe 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html
@@ -495,7 +495,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 private static HRegionLocation
-AsyncMetaTableAccessor.getRegionLocation(Resultr,
+MetaTableAccessor.getRegionLocation(Resultr,
  RegionInforegionInfo,
  intreplicaId)
 Returns the HRegionLocation parsed from the given meta row 
Result
@@ -504,7 +504,7 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 private static HRegionLocation
-MetaTableAccessor.getRegionLocation(Resultr,
+AsyncMetaTableAccessor.getRegionLocation(Resultr,
  RegionInforegionInfo,
  intreplicaId)
 Returns the HRegionLocation parsed from the given meta row 
Result
@@ -944,7 +944,9 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-AsyncHBaseAdmin.getRegions(ServerNameserverName)
+AsyncAdmin.getRegions(ServerNameserverName)
+Get all the online regions on a region server.
+
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
@@ -953,22 +955,22 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-HBaseAdmin.getRegions(ServerNamesn)
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
+RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-AsyncAdmin.getRegions(ServerNameserverName)
-Get all the online regions on a region server.
-
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
+HBaseAdmin.getRegions(ServerNamesn)
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-RawAsyncHBaseAdmin.getRegions(ServerNameserverName)
+AsyncHBaseAdmin.getRegions(ServerNameserverName)
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-AsyncHBaseAdmin.getRegions(TableNametableName)
+AsyncAdmin.getRegions(TableNametableName)
+Get the regions of a given table.
+
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
@@ -977,18 +979,16 @@ Input/OutputFormats, a table indexing MapReduce job, and 
utility methods.
 
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
-HBaseAdmin.getRegions(TableNametableName)
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
+RawAsyncHBaseAdmin.getRegions(TableNametableName)
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in 

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
index ad601c4..53e455f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html
@@ -1117,1183 +1117,1186 @@
 1109  @Nullable
 1110  public static TableState 
getTableState(Connection conn, TableName tableName)
   throws IOException {
-1112Table metaHTable = 
getMetaHTable(conn);
-1113Get get = new 
Get(tableName.getName()).addColumn(getTableFamily(), getTableStateColumn());
-1114long time = 
EnvironmentEdgeManager.currentTime();
-1115get.setTimeRange(0, time);
-1116Result result =
-1117metaHTable.get(get);
-1118return getTableState(result);
-1119  }
-1120
-1121  /**
-1122   * Fetch table states from META 
table
-1123   * @param conn connection to use
-1124   * @return map {tableName -gt; 
state}
-1125   * @throws IOException
-1126   */
-1127  public static MapTableName, 
TableState getTableStates(Connection conn)
-1128  throws IOException {
-1129final MapTableName, 
TableState states = new LinkedHashMap();
-1130Visitor collector = new Visitor() 
{
-1131  @Override
-1132  public boolean visit(Result r) 
throws IOException {
-1133TableState state = 
getTableState(r);
-1134if (state != null)
-1135  
states.put(state.getTableName(), state);
-1136return true;
-1137  }
-1138};
-1139fullScanTables(conn, collector);
-1140return states;
-1141  }
-1142
-1143  /**
-1144   * Updates state in META
-1145   * @param conn connection to use
-1146   * @param tableName table to look 
for
-1147   * @throws IOException
-1148   */
-1149  public static void 
updateTableState(Connection conn, TableName tableName,
-1150  TableState.State actual) throws 
IOException {
-1151updateTableState(conn, new 
TableState(tableName, actual));
-1152  }
-1153
-1154  /**
-1155   * Decode table state from META 
Result.
-1156   * Should contain cell from 
HConstants.TABLE_FAMILY
-1157   * @param r result
-1158   * @return null if not found
-1159   * @throws IOException
-1160   */
-1161  @Nullable
-1162  public static TableState 
getTableState(Result r)
-1163  throws IOException {
-1164Cell cell = 
r.getColumnLatestCell(getTableFamily(), getTableStateColumn());
-1165if (cell == null) return null;
-1166try {
-1167  return 
TableState.parseFrom(TableName.valueOf(r.getRow()),
-1168  
Arrays.copyOfRange(cell.getValueArray(),
-1169  cell.getValueOffset(), 
cell.getValueOffset() + cell.getValueLength()));
-1170} catch (DeserializationException e) 
{
-1171  throw new IOException(e);
-1172}
-1173
-1174  }
-1175
-1176  /**
-1177   * Implementations 'visit' a catalog 
table row.
-1178   */
-1179  public interface Visitor {
-1180/**
-1181 * Visit the catalog table row.
-1182 * @param r A row from catalog 
table
-1183 * @return True if we are to proceed 
scanning the table, else false if
-1184 * we are to stop now.
-1185 */
-1186boolean visit(final Result r) throws 
IOException;
-1187  }
-1188
-1189  /**
-1190   * Implementations 'visit' a catalog 
table row but with close() at the end.
-1191   */
-1192  public interface CloseableVisitor 
extends Visitor, Closeable {
-1193  }
-1194
-1195  /**
-1196   * A {@link Visitor} that collects 
content out of passed {@link Result}.
-1197   */
-1198  static abstract class 
CollectingVisitorT implements Visitor {
-1199final ListT results = new 
ArrayList();
-1200@Override
-1201public boolean visit(Result r) 
throws IOException {
-1202  if (r ==  null || r.isEmpty()) 
return true;
-1203  add(r);
-1204  return true;
-1205}
-1206
-1207abstract void add(Result r);
-1208
-1209/**
-1210 * @return Collected results; wait 
till visits complete to collect all
-1211 * possible results
-1212 */
-1213ListT getResults() {
-1214  return this.results;
-1215}
-1216  }
-1217
-1218  /**
-1219   * Collects all returned.
-1220   */
-1221  static class CollectAllVisitor extends 
CollectingVisitorResult {
-1222@Override
-1223void add(Result r) {
-1224  this.results.add(r);
-1225}
-1226  }
-1227
-1228  /**
-1229   * A Visitor that skips offline 
regions and split parents
-1230   */
-1231  public static abstract class 
DefaultVisitorBase implements Visitor {
-1232
-1233public DefaultVisitorBase() {
-1234  super();
-1235}
-1236
-1237public abstract boolean 
visitInternal(Result rowResult) throws IOException;
-1238
-1239

[27/51] [partial] hbase-site git commit: Published site at .

2018-02-02 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html 
b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
index ea234c2..34c375d 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html
@@ -567,7 +567,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 warningMap
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in 
java.util">Maporg.apache.hadoop.fs.FileSystem,http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean warningMap
+private static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in 
java.util">Maporg.apache.hadoop.fs.FileSystem,http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean warningMap
 
 
 
@@ -972,7 +972,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 isValidWALRootDir
-private staticbooleanisValidWALRootDir(org.apache.hadoop.fs.PathwalDir,
+private staticbooleanisValidWALRootDir(org.apache.hadoop.fs.PathwalDir,
  
org.apache.hadoop.conf.Configurationc)
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
@@ -987,7 +987,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getTableDir
-public staticorg.apache.hadoop.fs.PathgetTableDir(org.apache.hadoop.fs.Pathrootdir,
+public staticorg.apache.hadoop.fs.PathgetTableDir(org.apache.hadoop.fs.Pathrootdir,
 TableNametableName)
 Returns the Path object representing the table 
directory under
  path rootdir
@@ -1006,7 +1006,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getTableName
-public staticTableNamegetTableName(org.apache.hadoop.fs.PathtablePath)
+public staticTableNamegetTableName(org.apache.hadoop.fs.PathtablePath)
 Returns the TableName object representing
  the table directory under
  path rootdir
@@ -1024,7 +1024,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getNamespaceDir
-public staticorg.apache.hadoop.fs.PathgetNamespaceDir(org.apache.hadoop.fs.Pathrootdir,
+public staticorg.apache.hadoop.fs.PathgetNamespaceDir(org.apache.hadoop.fs.Pathrootdir,
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringnamespace)
 Returns the Path object representing
  the namespace directory under path rootdir
@@ -1043,7 +1043,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 setStoragePolicy
-public staticvoidsetStoragePolicy(org.apache.hadoop.fs.FileSystemfs,
+public staticvoidsetStoragePolicy(org.apache.hadoop.fs.FileSystemfs,
 
org.apache.hadoop.conf.Configurationconf,
 org.apache.hadoop.fs.Pathpath,
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringpolicyKey,
@@ -1074,7 +1074,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 setStoragePolicy
-public staticvoidsetStoragePolicy(org.apache.hadoop.fs.FileSystemfs,
+public staticvoidsetStoragePolicy(org.apache.hadoop.fs.FileSystemfs,
 org.apache.hadoop.fs.Pathpath,
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringstoragePolicy)
 Sets storage policy for given path.
@@ -1101,7 +1101,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 invokeSetStoragePolicy
-private staticvoidinvokeSetStoragePolicy(org.apache.hadoop.fs.FileSystemfs,
+private staticvoidinvokeSetStoragePolicy(org.apache.hadoop.fs.FileSystemfs,
org.apache.hadoop.fs.Pathpath,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringstoragePolicy)
 
@@ -1112,7 +1112,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 isHDFS
-public staticbooleanisHDFS(org.apache.hadoop.conf.Configurationconf)
+public 

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
 
b/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
index 633812a..d239905 100644
--- 
a/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html
@@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public class TestServerSideScanMetricsFromClientSide
+public class TestServerSideScanMetricsFromClientSide
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 
 
@@ -134,64 +134,68 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 CELL_HEAP_SIZE
 
 
+static HBaseClassTestRule
+CLASS_RULE
+
+
 private static byte[][]
 FAMILIES
 
-
+
 private static byte[]
 FAMILY
 
-
+
 private static int
 NUM_COLS
 
-
+
 private static int
 NUM_FAMILIES
 
-
+
 private static int
 NUM_QUALIFIERS
 
-
+
 private static int
 NUM_ROWS
 
-
+
 private static byte[]
 QUALIFIER
 
-
+
 private static byte[][]
 QUALIFIERS
 
-
+
 private static byte[]
 ROW
 
-
+
 private static byte[][]
 ROWS
 
-
+
 private static 
org.apache.hadoop.hbase.client.Table
 TABLE
 
-
+
 private static 
org.apache.hadoop.hbase.TableName
 TABLE_NAME
 Table configuration
 
 
-
+
 private static HBaseTestingUtility
 TEST_UTIL
 
-
+
 private static byte[]
 VALUE
 
-
+
 private static int
 VALUE_SIZE
 
@@ -316,13 +320,22 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 Field Detail
+
+
+
+
+
+CLASS_RULE
+public static finalHBaseClassTestRule CLASS_RULE
+
+
 
 
 
 
 
 TEST_UTIL
-private static finalHBaseTestingUtility TEST_UTIL
+private static finalHBaseTestingUtility TEST_UTIL
 
 
 
@@ -331,7 +344,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 TABLE
-private staticorg.apache.hadoop.hbase.client.Table TABLE
+private staticorg.apache.hadoop.hbase.client.Table TABLE
 
 
 
@@ -340,7 +353,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 TABLE_NAME
-private staticorg.apache.hadoop.hbase.TableName TABLE_NAME
+private staticorg.apache.hadoop.hbase.TableName TABLE_NAME
 Table configuration
 
 
@@ -350,7 +363,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 NUM_ROWS
-private staticint NUM_ROWS
+private staticint NUM_ROWS
 
 
 
@@ -359,7 +372,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 ROW
-private staticbyte[] ROW
+private staticbyte[] ROW
 
 
 
@@ -368,7 +381,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 ROWS
-private staticbyte[][] ROWS
+private staticbyte[][] ROWS
 
 
 
@@ -377,7 +390,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 NUM_FAMILIES
-private staticint NUM_FAMILIES
+private staticint NUM_FAMILIES
 
 
 
@@ -386,7 +399,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 FAMILY
-private staticbyte[] FAMILY
+private staticbyte[] FAMILY
 
 
 
@@ -395,7 +408,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 FAMILIES
-private staticbyte[][] FAMILIES
+private staticbyte[][] FAMILIES
 
 
 
@@ -404,7 +417,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 NUM_QUALIFIERS
-private staticint NUM_QUALIFIERS
+private staticint NUM_QUALIFIERS
 
 
 
@@ -413,7 +426,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 QUALIFIER
-private staticbyte[] QUALIFIER
+private staticbyte[] QUALIFIER
 
 
 
@@ -422,7 +435,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 QUALIFIERS
-private staticbyte[][] QUALIFIERS
+private staticbyte[][] QUALIFIERS
 
 
 
@@ -431,7 +444,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 VALUE_SIZE
-private staticint VALUE_SIZE
+private staticint VALUE_SIZE
 
 
 
@@ -440,7 +453,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 VALUE
-private staticbyte[] VALUE
+private staticbyte[] VALUE
 
 
 
@@ -449,7 +462,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 NUM_COLS
-private staticint NUM_COLS
+private staticint NUM_COLS
 
 
 
@@ -458,7 +471,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 CELL_HEAP_SIZE
-private staticlong CELL_HEAP_SIZE
+private staticlong CELL_HEAP_SIZE
 
 
 
@@ -475,7 +488,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 TestServerSideScanMetricsFromClientSide
-publicTestServerSideScanMetricsFromClientSide()

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
index eb9e252..667152a 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
@@ -28,22 +28,22 @@
 020
 021import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH;
 022import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC;
-023import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
-024import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
-025import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
-026import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
-027import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
-028import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
-029import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
-030import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
-031import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
-032import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
-033import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
-034import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
-035import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
-036import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
-037import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
-038import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
+023import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG;
+024import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC;
+025import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH;
+026import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC;
+027import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER;
+028import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC;
+029import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET;
+030import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC;
+031import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC;
+032import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE;
+033import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC;
+034import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC;
+035import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS;
+036import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC;
+037import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME;
+038import static 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC;
 039
 040import java.io.IOException;
 041import java.net.URI;
@@ -70,194 +70,194 @@
 062import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
 063import 
org.apache.hadoop.hbase.client.Connection;
 064import 
org.apache.hadoop.hbase.client.ConnectionFactory;
-065import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
-066import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-067import 
org.apache.yetus.audience.InterfaceAudience;
-068
-069/**
-070 * General backup commands, options and 
usage messages
-071 */
-072
+065import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+066import 
org.apache.yetus.audience.InterfaceAudience;
+067
+068import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
+069
+070/**
+071 * General backup commands, options and 
usage messages
+072 */
 073@InterfaceAudience.Private
 074public final class BackupCommands {
-075
-076  public final static String 
INCORRECT_USAGE = "Incorrect usage";
-077
-078  public final static String 
TOP_LEVEL_NOT_ALLOWED =
-079  "Top level (root) folder is not 
allowed to be a backup destination";
-080
-081  public static final String USAGE = 
"Usage: hbase 

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
index 0164ad3..fd16dee 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html
@@ -76,64 +76,54 @@
 068  }
 069
 070  @Override
-071  void setTimeProgress(long timeProgress) 
{
+071  void setProgress(int batchProgress, 
long sizeProgress, long heapSizeProgress) {
 072// Do nothing. NoLimitScannerContext 
instances are immutable post-construction
 073  }
 074
 075  @Override
-076  void updateTimeProgress() {
+076  void clearProgress() {
 077// Do nothing. NoLimitScannerContext 
instances are immutable post-construction
 078  }
 079
 080  @Override
-081  void setProgress(int batchProgress, 
long sizeProgress, long heapSizeProgress, long timeProgress) {
+081  void setSizeLimitScope(LimitScope 
scope) {
 082// Do nothing. NoLimitScannerContext 
instances are immutable post-construction
 083  }
 084
 085  @Override
-086  void clearProgress() {
+086  void setTimeLimitScope(LimitScope 
scope) {
 087// Do nothing. NoLimitScannerContext 
instances are immutable post-construction
 088  }
 089
 090  @Override
-091  void setSizeLimitScope(LimitScope 
scope) {
+091  NextState setScannerState(NextState 
state) {
 092// Do nothing. NoLimitScannerContext 
instances are immutable post-construction
-093  }
-094
-095  @Override
-096  void setTimeLimitScope(LimitScope 
scope) {
-097// Do nothing. NoLimitScannerContext 
instances are immutable post-construction
-098  }
-099
-100  @Override
-101  NextState setScannerState(NextState 
state) {
-102// Do nothing. NoLimitScannerContext 
instances are immutable post-construction
-103return state;
-104  }
-105
-106  @Override
-107  boolean checkBatchLimit(LimitScope 
checkerScope) {
-108// No limits can be specified, thus 
return false to indicate no limit has been reached.
-109return false;
-110  }
-111
-112  @Override
-113  boolean checkSizeLimit(LimitScope 
checkerScope) {
-114// No limits can be specified, thus 
return false to indicate no limit has been reached.
-115return false;
-116  }
-117
-118  @Override
-119  boolean checkTimeLimit(LimitScope 
checkerScope) {
-120// No limits can be specified, thus 
return false to indicate no limit has been reached.
-121return false;
-122  }
-123
-124  @Override
-125  boolean checkAnyLimitReached(LimitScope 
checkerScope) {
-126return false;
-127  }
-128}
+093return state;
+094  }
+095
+096  @Override
+097  boolean checkBatchLimit(LimitScope 
checkerScope) {
+098// No limits can be specified, thus 
return false to indicate no limit has been reached.
+099return false;
+100  }
+101
+102  @Override
+103  boolean checkSizeLimit(LimitScope 
checkerScope) {
+104// No limits can be specified, thus 
return false to indicate no limit has been reached.
+105return false;
+106  }
+107
+108  @Override
+109  boolean checkTimeLimit(LimitScope 
checkerScope) {
+110// No limits can be specified, thus 
return false to indicate no limit has been reached.
+111return false;
+112  }
+113
+114  @Override
+115  boolean checkAnyLimitReached(LimitScope 
checkerScope) {
+116return false;
+117  }
+118}
 
 
 



[27/51] [partial] hbase-site git commit: Published site at .

2018-01-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html 
b/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html
index f8c49cb..e3762b3 100644
--- a/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html
+++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html
@@ -1265,7 +1265,7 @@ implements 
 
 CACHE_FIXED_OVERHEAD
-public static finallong CACHE_FIXED_OVERHEAD
+public static finallong CACHE_FIXED_OVERHEAD
 
 
 
@@ -1434,7 +1434,7 @@ implements 
 
 cacheBlock
-publicvoidcacheBlock(BlockCacheKeycacheKey,
+publicvoidcacheBlock(BlockCacheKeycacheKey,
Cacheablebuf)
 Cache the block with the specified name and buffer.
  
@@ -1453,7 +1453,7 @@ implements 
 
 updateSizeMetrics
-privatelongupdateSizeMetrics(LruCachedBlockcb,
+privatelongupdateSizeMetrics(LruCachedBlockcb,
booleanevict)
 Helper function that updates the local size counter and 
also updates any
  per-cf or per-blocktype metrics it can discern from given
@@ -1466,7 +1466,7 @@ implements 
 
 getBlock
-publicCacheablegetBlock(BlockCacheKeycacheKey,
+publicCacheablegetBlock(BlockCacheKeycacheKey,
   booleancaching,
   booleanrepeat,
   booleanupdateCacheMetrics)
@@ -1492,7 +1492,7 @@ implements 
 
 containsBlock
-publicbooleancontainsBlock(BlockCacheKeycacheKey)
+publicbooleancontainsBlock(BlockCacheKeycacheKey)
 Whether the cache contains block with specified 
cacheKey
 
 Returns:
@@ -1506,7 +1506,7 @@ implements 
 
 evictBlock
-publicbooleanevictBlock(BlockCacheKeycacheKey)
+publicbooleanevictBlock(BlockCacheKeycacheKey)
 Description copied from 
interface:BlockCache
 Evict block from cache.
 
@@ -1525,7 +1525,7 @@ implements 
 
 evictBlocksByHfileName
-publicintevictBlocksByHfileName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringhfileName)
+publicintevictBlocksByHfileName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringhfileName)
 Evicts all blocks for a specific HFile. This is an
  expensive operation implemented as a linear-time search through all blocks
  in the cache. Ideally this should be a search in a log-access-time map.
@@ -1546,7 +1546,7 @@ implements 
 
 evictBlock
-protectedlongevictBlock(LruCachedBlockblock,
+protectedlongevictBlock(LruCachedBlockblock,
   booleanevictedByEvictionProcess)
 Evict the block, and it will be cached by the victim 
handler if exists 
  block may be read again later
@@ -1565,7 +1565,7 @@ implements 
 
 runEviction
-privatevoidrunEviction()
+privatevoidrunEviction()
 Multi-threaded call to run the eviction process.
 
 
@@ -1575,7 +1575,7 @@ implements 
 
 isEvictionInProgress
-booleanisEvictionInProgress()
+booleanisEvictionInProgress()
 
 
 
@@ -1584,7 +1584,7 @@ implements 
 
 getOverhead
-longgetOverhead()
+longgetOverhead()
 
 
 
@@ -1593,7 +1593,7 @@ implements 
 
 evict
-voidevict()
+voidevict()
 Eviction method.
 
 
@@ -1603,7 +1603,7 @@ implements 
 
 toString
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtoString()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtoString()
 
 Overrides:
 http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
 title="class or interface in java.lang">toStringin 
classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
@@ -1616,7 +1616,7 @@ implements 
 
 getMaxSize
-publiclonggetMaxSize()
+publiclonggetMaxSize()
 Get the maximum size of this cache.
 
 Specified by:
@@ -1632,7 +1632,7 @@ implements 
 
 getCurrentSize
-publiclonggetCurrentSize()
+publiclonggetCurrentSize()
 Description copied from 
interface:BlockCache
 Returns the occupied size of the block cache, in 
bytes.
 
@@ -1649,7 +1649,7 @@ implements 
 
 getCurrentDataSize
-publiclonggetCurrentDataSize()
+publiclonggetCurrentDataSize()
 Description copied from 
interface:BlockCache
 Returns the occupied size of data blocks, in bytes.
 
@@ -1666,7 +1666,7 @@ implements 
 
 getFreeSize
-publiclonggetFreeSize()
+publiclonggetFreeSize()
 Description copied from 
interface:BlockCache
 Returns the free size of the block cache, in bytes.
 
@@ -1683,7 +1683,7 @@ implements 
 
 size
-publiclongsize()
+publiclongsize()
 Description copied from 
interface:BlockCache
 Returns the total size of the block cache, in bytes.
 
@@ -1700,7 +1700,7 @@ implements 
 
 getBlockCount

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
index e05510e..2e114d8 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html
@@ -54,891 +54,884 @@
 046import 
org.apache.hadoop.hbase.HTableDescriptor;
 047import 
org.apache.hadoop.hbase.MetaTableAccessor;
 048import 
org.apache.hadoop.hbase.MetaTableAccessor.DefaultVisitorBase;
-049import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-050import 
org.apache.hadoop.hbase.ServerName;
-051import 
org.apache.hadoop.hbase.TableName;
-052import 
org.apache.hadoop.hbase.client.ClusterConnection;
-053import 
org.apache.hadoop.hbase.client.Delete;
-054import 
org.apache.hadoop.hbase.client.Get;
-055import 
org.apache.hadoop.hbase.client.Mutation;
-056import 
org.apache.hadoop.hbase.client.Put;
-057import 
org.apache.hadoop.hbase.client.RegionInfo;
-058import 
org.apache.hadoop.hbase.client.Result;
-059import 
org.apache.hadoop.hbase.client.Scan;
-060import 
org.apache.hadoop.hbase.client.Table;
-061import 
org.apache.hadoop.hbase.client.TableState;
-062import 
org.apache.hadoop.hbase.constraint.ConstraintException;
-063import 
org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint;
-064import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-065import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
-066import 
org.apache.hadoop.hbase.master.MasterServices;
-067import 
org.apache.hadoop.hbase.master.ServerListener;
-068import 
org.apache.hadoop.hbase.master.TableStateManager;
-069import 
org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode;
-070import 
org.apache.hadoop.hbase.net.Address;
-071import 
org.apache.hadoop.hbase.procedure2.Procedure;
-072import 
org.apache.hadoop.hbase.protobuf.ProtobufMagic;
-073import 
org.apache.hadoop.hbase.protobuf.ProtobufUtil;
-074import 
org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos;
-075import 
org.apache.hadoop.hbase.protobuf.generated.RSGroupProtos;
-076import 
org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy;
-077import 
org.apache.hadoop.hbase.security.access.AccessControlLists;
-078import 
org.apache.hadoop.hbase.util.Bytes;
-079import 
org.apache.hadoop.hbase.zookeeper.ZKUtil;
-080import 
org.apache.hadoop.hbase.zookeeper.ZKWatcher;
-081import 
org.apache.hadoop.hbase.zookeeper.ZNodePaths;
-082import 
org.apache.yetus.audience.InterfaceAudience;
-083import 
org.apache.zookeeper.KeeperException;
-084import org.slf4j.Logger;
-085import org.slf4j.LoggerFactory;
-086
-087import 
org.apache.hbase.thirdparty.com.google.common.collect.Lists;
-088import 
org.apache.hbase.thirdparty.com.google.common.collect.Maps;
-089import 
org.apache.hbase.thirdparty.com.google.common.collect.Sets;
-090import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-091import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-092
-093/**
-094 * This is an implementation of {@link 
RSGroupInfoManager} which makes
-095 * use of an HBase table as the 
persistence store for the group information.
-096 * It also makes use of zookeeper to 
store group information needed
-097 * for bootstrapping during offline 
mode.
-098 *
-099 * h2Concurrency/h2
-100 * RSGroup state is kept locally in Maps. 
There is a rsgroup name to cached
-101 * RSGroupInfo Map at {@link #rsGroupMap} 
and a Map of tables to the name of the
-102 * rsgroup they belong too (in {@link 
#tableMap}). These Maps are persisted to the
-103 * hbase:rsgroup table (and cached in zk) 
on each modification.
-104 *
-105 * pMutations on state are 
synchronized but reads can continue without having
-106 * to wait on an instance monitor, 
mutations do wholesale replace of the Maps on
-107 * update -- Copy-On-Write; the local 
Maps of state are read-only, just-in-case
-108 * (see flushConfig).
-109 *
-110 * pReads must not block else 
there is a danger we'll deadlock.
-111 *
-112 * pClients of this class, the 
{@link RSGroupAdminEndpoint} for example, want to query and
-113 * then act on the results of the query 
modifying cache in zookeeper without another thread
-114 * making intermediate modifications. 
These clients synchronize on the 'this' instance so
-115 * no other has access concurrently. 
Reads must be able to continue concurrently.
-116 */
-117@InterfaceAudience.Private
-118final class RSGroupInfoManagerImpl 
implements RSGroupInfoManager {
-119  private static final Logger LOG = 

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
index bf1e334..f55c3a6 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-client archetype  Reactor 
Dependency Convergence
 
@@ -912,7 +912,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-01-11
+  Last Published: 
2018-01-12
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
index bdccdc9..692a5be 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-client archetype  
Dependency Information
 
@@ -147,7 +147,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-01-11
+  Last Published: 
2018-01-12
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
index 482ab8a..04acfbe 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-client archetype  Project 
Dependency Management
 
@@ -810,7 +810,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-01-11
+  Last Published: 
2018-01-12
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
index 9bf6e8e..9d5eee1 100644
--- a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
+++ b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-client archetype  
About
 
@@ -119,7 +119,7 @@
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2018-01-11
+  Last Published: 
2018-01-12
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
index 06a6c6f..3b2cf47 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-client archetype  CI 
Management
 
@@ -126,7 +126,7 @@
 

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-11 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
index 0b621b6..69a8636 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html
@@ -82,7 +82,7 @@
 074  }
 075
 076  @Override
-077  public WAL getWAL(final byte[] 
identifier, byte[] namespace) throws IOException {
+077  public WAL getWAL(RegionInfo region) 
throws IOException {
 078return disabled;
 079  }
 080

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html
index 0b621b6..69a8636 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html
@@ -82,7 +82,7 @@
 074  }
 075
 076  @Override
-077  public WAL getWAL(final byte[] 
identifier, byte[] namespace) throws IOException {
+077  public WAL getWAL(RegionInfo region) 
throws IOException {
 078return disabled;
 079  }
 080

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
index 4fec5eb..63c9ca7 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html
@@ -35,254 +35,262 @@
 027import java.util.List;
 028import 
java.util.concurrent.ConcurrentHashMap;
 029import 
java.util.concurrent.ConcurrentMap;
-030
+030import java.util.concurrent.locks.Lock;
 031import 
org.apache.hadoop.conf.Configuration;
-032import 
org.apache.yetus.audience.InterfaceAudience;
-033import org.slf4j.Logger;
-034import org.slf4j.LoggerFactory;
-035// imports for classes still in 
regionserver.wal
-036import 
org.apache.hadoop.hbase.regionserver.wal.WALActionsListener;
-037import 
org.apache.hadoop.hbase.util.Bytes;
-038import 
org.apache.hadoop.hbase.util.IdLock;
-039
-040/**
-041 * A WAL Provider that returns a WAL per 
group of regions.
-042 *
-043 * This provider follows the decorator 
pattern and mainly holds the logic for WAL grouping.
-044 * WAL creation/roll/close is delegated 
to {@link #DELEGATE_PROVIDER}
-045 *
-046 * Region grouping is handled via {@link 
RegionGroupingStrategy} and can be configured via the
-047 * property 
"hbase.wal.regiongrouping.strategy". Current strategy choices are
-048 * ul
-049 *   
liemdefaultStrategy/em : Whatever strategy this version 
of HBase picks. currently
-050 *  
"bounded"./li
-051 *   
liemidentity/em : each region belongs to its own 
group./li
-052 *   
liembounded/em : bounded number of groups and region 
evenly assigned to each group./li
-053 * /ul
-054 * Optionally, a FQCN to a custom 
implementation may be given.
-055 */
-056@InterfaceAudience.Private
-057public class RegionGroupingProvider 
implements WALProvider {
-058  private static final Logger LOG = 
LoggerFactory.getLogger(RegionGroupingProvider.class);
-059
-060  /**
-061   * Map identifiers to a group number.
-062   */
-063  public static interface 
RegionGroupingStrategy {
-064String GROUP_NAME_DELIMITER = ".";
-065
-066/**
-067 * Given an identifier and a 
namespace, pick a group.
-068 */
-069String group(final byte[] identifier, 
byte[] namespace);
-070void init(Configuration config, 
String providerId);
-071  }
-072
-073  /**
-074   * Maps between configuration names for 
strategies and implementation classes.
-075   */
-076  static enum Strategies {
-077
defaultStrategy(BoundedGroupingStrategy.class),
-078
identity(IdentityGroupingStrategy.class),
-079
bounded(BoundedGroupingStrategy.class),
-080
namespace(NamespaceGroupingStrategy.class);
-081
-082final Class? extends 
RegionGroupingStrategy clazz;
-083Strategies(Class? extends 
RegionGroupingStrategy clazz) {
-084  this.clazz = clazz;
-085}
-086  }
-087
-088  /**
-089   * instantiate a strategy from a config 

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
index 67417b5..900808a 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HBaseAdmin.AbortProcedureFuture
+private static class HBaseAdmin.AbortProcedureFuture
 extends HBaseAdmin.ProcedureFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean
 
 
@@ -235,7 +235,7 @@ extends 
 
 isAbortInProgress
-privateboolean isAbortInProgress
+privateboolean isAbortInProgress
 
 
 
@@ -252,7 +252,7 @@ extends 
 
 AbortProcedureFuture
-publicAbortProcedureFuture(HBaseAdminadmin,
+publicAbortProcedureFuture(HBaseAdminadmin,
 http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">LongprocId,
 http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in 
java.lang">BooleanabortProcResponse)
 
@@ -271,7 +271,7 @@ extends 
 
 get
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Booleanget(longtimeout,
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Booleanget(longtimeout,
http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true;
 title="class or interface in java.util.concurrent">TimeUnitunit)
 throws http://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
 title="class or interface in java.lang">InterruptedException,
http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutionException.html?is-external=true;
 title="class or interface in java.util.concurrent">ExecutionException,

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
index 4e91404..9d834f8 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html
@@ -132,7 +132,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HBaseAdmin.AddColumnFamilyFuture
+private static class HBaseAdmin.AddColumnFamilyFuture
 extends HBaseAdmin.ModifyTableFuture
 
 
@@ -246,7 +246,7 @@ extends 
 
 AddColumnFamilyFuture
-publicAddColumnFamilyFuture(HBaseAdminadmin,
+publicAddColumnFamilyFuture(HBaseAdminadmin,
  TableNametableName,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponseresponse)
 
@@ -265,7 +265,7 @@ extends 
 
 getOperationType
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetOperationType()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetOperationType()
 
 Overrides:
 getOperationTypein
 classHBaseAdmin.ModifyTableFuture

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
index 485fa7a..40023f1 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html
@@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HBaseAdmin.CreateTableFuture
+private static class HBaseAdmin.CreateTableFuture
 extends HBaseAdmin.TableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
 
 
@@ -268,7 +268,7 @@ extends 
 
 desc
-private finalTableDescriptor desc
+private 

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-04 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
index 19aa0b9..aedcf26 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-class RawAsyncHBaseAdmin
+class RawAsyncHBaseAdmin
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements AsyncAdmin
 The implementation of AsyncAdmin.
@@ -704,12 +704,12 @@ implements 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureClusterStatus
-getClusterStatus()
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics
+getClusterMetrics()
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureClusterStatus
-getClusterStatus(http://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true;
 title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions)
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics
+getClusterMetrics(http://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true;
 title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions)
 
 
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureCompactionState
@@ -788,27 +788,27 @@ implements 
-private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionLoad
-getRegionLoad(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequestrequest,
- ServerNameserverName)
+(package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFutureHRegionLocation
+getRegionLocation(byte[]regionNameOrEncodedRegionName)
+Get the region location for the passed region name.
+
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionLoad
-getRegionLoads(ServerNameserverName)
-Get a list of RegionLoad of all regions hosted on a 
region seerver.
-
+private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionMetrics
+getRegionMetrics(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequestrequest,
+ServerNameserverName)
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionLoad
-getRegionLoads(ServerNameserverName,
-  TableNametableName)
-Get a list of RegionLoad of all regions hosted on a 
region seerver for a table.
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionMetrics
+getRegionMetrics(ServerNameserverName)
+Get a list of RegionMetrics of 
all regions hosted on a region seerver.
 
 
 
-(package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in 

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
index 5b3b750..a1f3f7e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html
@@ -97,3307 +97,3304 @@
 089import 
org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer;
 090import 
org.apache.hbase.thirdparty.io.netty.util.Timeout;
 091import 
org.apache.hbase.thirdparty.io.netty.util.TimerTask;
-092import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-093import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-094import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-095import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
-096import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
-098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse;
-099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
-105import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest;
-106import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse;
-107import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest;
-108import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse;
-109import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest;
-110import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse;
-111import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest;
-112import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse;
-113import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
-114import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
-115import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-116import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
-117import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
-118import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-119import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-120import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-121import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
-122import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
-123import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
-124import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest;
-125import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse;
-126import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-127import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-128import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest;
-129import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse;
-130import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest;
-131import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse;
-132import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
-133import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;

[27/51] [partial] hbase-site git commit: Published site at .

2018-01-01 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
 
b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
index 1fd770d..5c0a3d9 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html
@@ -1961,6 +1961,6 @@ publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.h
 
 
 
-Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html 
b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html
index 1ce7c03..782e4e3 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html
@@ -1553,6 +1553,6 @@ publicCopyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html 
b/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
index 037cb4a..dd4afb1 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html
@@ -620,6 +620,6 @@ extends Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html 
b/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
index f65cdb2..66f3198 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html
@@ -409,6 +409,6 @@ not permitted.)
 
 
 
-Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableState.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableState.html 
b/devapidocs/org/apache/hadoop/hbase/client/TableState.html
index 310a067..0ea4ab9 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/TableState.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/TableState.html
@@ -556,6 +556,6 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 
-Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html 
b/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html
index 76cec51..6e021bc 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html
@@ -356,6 +356,6 @@ extends 
 
 
-Copyright  20072017 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
+Copyright  20072018 https://www.apache.org/;>The Apache Software Foundation. All rights 
reserved.
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/VersionInfoUtil.html

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-30 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
index 91eec45..d1cd185 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html
@@ -88,400 +88,396 @@
 080  private final static String 
CONF_COMPACT_ONCE = "hbase.compactiontool.compact.once";
 081  private final static String 
CONF_COMPACT_MAJOR = "hbase.compactiontool.compact.major";
 082  private final static String 
CONF_DELETE_COMPACTED = "hbase.compactiontool.delete";
-083  private final static String 
CONF_COMPLETE_COMPACTION = "hbase.hstore.compaction.complete";
-084
-085  /**
-086   * Class responsible to execute the 
Compaction on the specified path.
-087   * The path can be a table, region or 
family directory.
-088   */
-089  private static class CompactionWorker 
{
-090private final boolean 
keepCompactedFiles;
-091private final boolean 
deleteCompacted;
-092private final Configuration conf;
-093private final FileSystem fs;
-094private final Path tmpDir;
-095
-096public CompactionWorker(final 
FileSystem fs, final Configuration conf) {
-097  this.conf = conf;
-098  this.keepCompactedFiles = 
!conf.getBoolean(CONF_COMPLETE_COMPACTION, true);
-099  this.deleteCompacted = 
conf.getBoolean(CONF_DELETE_COMPACTED, false);
-100  this.tmpDir = new 
Path(conf.get(CONF_TMP_DIR));
-101  this.fs = fs;
-102}
-103
-104/**
-105 * Execute the compaction on the 
specified path.
-106 *
-107 * @param path Directory path on 
which to run compaction.
-108 * @param compactOnce Execute just a 
single step of compaction.
-109 * @param major Request major 
compaction.
-110 */
-111public void compact(final Path path, 
final boolean compactOnce, final boolean major) throws IOException {
-112  if (isFamilyDir(fs, path)) {
-113Path regionDir = 
path.getParent();
-114Path tableDir = 
regionDir.getParent();
-115TableDescriptor htd = 
FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir);
-116RegionInfo hri = 
HRegionFileSystem.loadRegionInfoFileContent(fs, regionDir);
-117compactStoreFiles(tableDir, htd, 
hri,
-118path.getName(), compactOnce, 
major);
-119  } else if (isRegionDir(fs, path)) 
{
-120Path tableDir = 
path.getParent();
-121TableDescriptor htd = 
FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir);
-122compactRegion(tableDir, htd, 
path, compactOnce, major);
-123  } else if (isTableDir(fs, path)) 
{
-124compactTable(path, compactOnce, 
major);
-125  } else {
-126throw new IOException(
-127  "Specified path is not a table, 
region or family directory. path=" + path);
-128  }
-129}
-130
-131private void compactTable(final Path 
tableDir, final boolean compactOnce, final boolean major)
-132throws IOException {
-133  TableDescriptor htd = 
FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir);
-134  for (Path regionDir: 
FSUtils.getRegionDirs(fs, tableDir)) {
-135compactRegion(tableDir, htd, 
regionDir, compactOnce, major);
-136  }
-137}
-138
-139private void compactRegion(final Path 
tableDir, final TableDescriptor htd,
-140final Path regionDir, final 
boolean compactOnce, final boolean major)
-141throws IOException {
-142  RegionInfo hri = 
HRegionFileSystem.loadRegionInfoFileContent(fs, regionDir);
-143  for (Path familyDir: 
FSUtils.getFamilyDirs(fs, regionDir)) {
-144compactStoreFiles(tableDir, htd, 
hri, familyDir.getName(), compactOnce, major);
-145  }
-146}
-147
-148/**
-149 * Execute the actual compaction 
job.
-150 * If the compact once flag is not 
specified, execute the compaction until
-151 * no more compactions are needed. 
Uses the Configuration settings provided.
-152 */
-153private void compactStoreFiles(final 
Path tableDir, final TableDescriptor htd,
-154final RegionInfo hri, final 
String familyName, final boolean compactOnce,
-155final boolean major) throws 
IOException {
-156  HStore store = getStore(conf, fs, 
tableDir, htd, hri, familyName, tmpDir);
-157  LOG.info("Compact table=" + 
htd.getTableName() +
-158" region=" + 
hri.getRegionNameAsString() +
-159" family=" + familyName);
-160  if (major) {
-161store.triggerMajorCompaction();
-162  }
-163  do {
-164OptionalCompactionContext 
compaction =
-165

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html 
b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
index edd369a..6fa4fbd 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
@@ -175,32 +175,32 @@ implements 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse
-abortProcedure(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+abortProcedure(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,
   
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse
-addColumn(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+addColumn(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse
-addReplicationPeer(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+addReplicationPeer(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,
   
org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse
-assignRegion(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+assignRegion(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse
-balance(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+balance(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,

org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse
-clearDeadServers(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+clearDeadServers(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequestrequest)
 
 
@@ -209,347 +209,347 @@ implements 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse
-createNamespace(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+createNamespace(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,

org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse
-createTable(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+createTable(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,

org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse
-decommissionRegionServers(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+decommissionRegionServers(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse
-deleteColumn(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+deleteColumn(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse
-deleteNamespace(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+deleteNamespace(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller,

org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequestrequest)
 
 
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html
 
b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html
new file mode 100644
index 000..662952e
--- /dev/null
+++ 
b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html
@@ -0,0 +1,1117 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+PrivateCellUtil.TagRewriteByteBufferExtendedCell (Apache HBase 
3.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase
+Class PrivateCellUtil.TagRewriteByteBufferExtendedCell
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.ByteBufferExtendedCell
+
+
+org.apache.hadoop.hbase.PrivateCellUtil.TagRewriteByteBufferExtendedCell
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true;
 title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell
+
+
+Direct Known Subclasses:
+PrivateCellUtil.ValueAndTagRewriteByteBufferExtendedCell
+
+
+Enclosing class:
+PrivateCellUtil
+
+
+
+static class PrivateCellUtil.TagRewriteByteBufferExtendedCell
+extends ByteBufferExtendedCell
+
+
+
+
+
+
+
+
+
+
+
+Nested Class Summary
+
+
+
+
+Nested classes/interfaces inherited from 
interfaceorg.apache.hadoop.hbase.Cell
+Cell.Type
+
+
+
+
+
+
+
+
+Field Summary
+
+Fields
+
+Modifier and Type
+Field and Description
+
+
+protected ByteBufferExtendedCell
+cell
+
+
+private static int
+HEAP_SIZE_OVERHEAD
+
+
+protected byte[]
+tags
+
+
+
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell
+CELL_NOT_BASED_ON_CHUNK
+
+
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.RawCell
+MAX_TAGS_LENGTH
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+TagRewriteByteBufferExtendedCell(ByteBufferExtendedCellcell,
+byte[]tags)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsInstance MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+ExtendedCell
+deepClone()
+Does a deep copy of the contents to a new memory area and 
returns it as a new cell.
+
+
+
+byte[]
+getFamilyArray()
+Contiguous bytes composed of legal HDFS filename characters 
which may start at any index in the
+ containing array.
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
 title="class or interface in java.nio">ByteBuffer
+getFamilyByteBuffer()
+
+
+byte
+getFamilyLength()
+
+
+int
+getFamilyOffset()
+
+
+int
+getFamilyPosition()
+
+
+byte[]
+getQualifierArray()
+Contiguous raw bytes that may start at any index in the 
containing array.
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
 title="class or interface in java.nio">ByteBuffer
+getQualifierByteBuffer()
+
+
+int
+getQualifierLength()
+
+
+int
+getQualifierOffset()
+
+
+int
+getQualifierPosition()
+
+
+byte[]
+getRowArray()
+Contiguous raw bytes that may start at any index in the 
containing array.
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
 title="class or interface in java.nio">ByteBuffer
+getRowByteBuffer()
+
+
+short
+getRowLength()
+
+
+int
+getRowOffset()
+
+
+int
+getRowPosition()
+
+

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
 
b/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
index 3d37d41..5a14666 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html
@@ -132,7 +132,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static final class HttpProxyExample.RequestHandler
+private static final class HttpProxyExample.RequestHandler
 extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHandlerorg.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequest
 
 
@@ -302,7 +302,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 conn
-private finalAsyncConnection conn
+private finalAsyncConnection conn
 
 
 
@@ -311,7 +311,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 channelGroup
-private 
finalorg.apache.hadoop.hbase.shaded.io.netty.channel.group.ChannelGroup 
channelGroup
+private 
finalorg.apache.hadoop.hbase.shaded.io.netty.channel.group.ChannelGroup 
channelGroup
 
 
 
@@ -328,7 +328,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 RequestHandler
-publicRequestHandler(AsyncConnectionconn,
+publicRequestHandler(AsyncConnectionconn,
   
org.apache.hadoop.hbase.shaded.io.netty.channel.group.ChannelGroupchannelGroup)
 
 
@@ -346,7 +346,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 channelActive
-publicvoidchannelActive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx)
+publicvoidchannelActive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx)
 
 Specified by:
 channelActivein 
interfaceorg.apache.hadoop.hbase.shaded.io.netty.channel.ChannelInboundHandler
@@ -361,7 +361,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 channelInactive
-publicvoidchannelInactive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx)
+publicvoidchannelInactive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx)
 
 Specified by:
 channelInactivein 
interfaceorg.apache.hadoop.hbase.shaded.io.netty.channel.ChannelInboundHandler
@@ -376,7 +376,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 write
-privatevoidwrite(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx,
+privatevoidwrite(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx,

org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.HttpResponseStatusstatus,
http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">Optionalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringcontent)
 
@@ -387,7 +387,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 parse
-privateHttpProxyExample.Paramsparse(org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq)
+privateHttpProxyExample.Paramsparse(org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq)
 
 
 
@@ -396,7 +396,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 get
-privatevoidget(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx,
+privatevoidget(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx,
  
org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq)
 
 
@@ -406,7 +406,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 put
-privatevoidput(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx,
+privatevoidput(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx,
  
org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq)
 
 
@@ -416,7 +416,7 @@ extends 
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand
 
 
 channelRead0
-protectedvoidchannelRead0(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx,
+protectedvoidchannelRead0(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx,
 
org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq)
 
 Specified by:
@@ -430,7 +430,7 @@ extends 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html 
b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
index 0c57dc7..061085c 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
@@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.LimitedPrivate(value="Tools")
-public class HMaster
+public class HMaster
 extends HRegionServer
 implements MasterServices
 HMaster is the "master server" for HBase. An HBase cluster 
has one active
@@ -1434,7 +1434,7 @@ implements 
 
 LOG
-private staticorg.slf4j.Logger LOG
+private staticorg.slf4j.Logger LOG
 
 
 
@@ -1443,7 +1443,7 @@ implements 
 
 MASTER
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MASTER
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MASTER
 
 See Also:
 Constant
 Field Values
@@ -1456,7 +1456,7 @@ implements 
 
 activeMasterManager
-private finalActiveMasterManager activeMasterManager
+private finalActiveMasterManager activeMasterManager
 
 
 
@@ -1465,7 +1465,7 @@ implements 
 
 regionServerTracker
-RegionServerTracker regionServerTracker
+RegionServerTracker regionServerTracker
 
 
 
@@ -1474,7 +1474,7 @@ implements 
 
 drainingServerTracker
-privateDrainingServerTracker drainingServerTracker
+privateDrainingServerTracker drainingServerTracker
 
 
 
@@ -1483,7 +1483,7 @@ implements 
 
 loadBalancerTracker
-LoadBalancerTracker loadBalancerTracker
+LoadBalancerTracker loadBalancerTracker
 
 
 
@@ -1492,7 +1492,7 @@ implements 
 
 splitOrMergeTracker
-privateSplitOrMergeTracker splitOrMergeTracker
+privateSplitOrMergeTracker splitOrMergeTracker
 
 
 
@@ -1501,7 +1501,7 @@ implements 
 
 regionNormalizerTracker
-privateRegionNormalizerTracker 
regionNormalizerTracker
+privateRegionNormalizerTracker 
regionNormalizerTracker
 
 
 
@@ -1510,7 +1510,7 @@ implements 
 
 maintenanceModeTracker
-privateMasterMaintenanceModeTracker maintenanceModeTracker
+privateMasterMaintenanceModeTracker maintenanceModeTracker
 
 
 
@@ -1519,7 +1519,7 @@ implements 
 
 clusterSchemaService
-privateClusterSchemaService clusterSchemaService
+privateClusterSchemaService clusterSchemaService
 
 
 
@@ -1528,7 +1528,7 @@ implements 
 
 HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
 
 See Also:
 Constant
 Field Values
@@ -1541,7 +1541,7 @@ implements 
 
 DEFAULT_HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
-public static finalint DEFAULT_HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
+public static finalint DEFAULT_HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS
 
 See Also:
 Constant
 Field Values
@@ -1554,7 +1554,7 @@ implements 
 
 metricsMaster
-finalMetricsMaster metricsMaster
+finalMetricsMaster metricsMaster
 
 
 
@@ -1563,7 +1563,7 @@ implements 
 
 fileSystemManager
-privateMasterFileSystem fileSystemManager
+privateMasterFileSystem fileSystemManager
 
 
 
@@ -1572,7 +1572,7 @@ implements 
 
 walManager
-privateMasterWalManager walManager
+privateMasterWalManager walManager
 
 
 
@@ -1581,7 +1581,7 @@ implements 
 
 serverManager
-private volatileServerManager serverManager
+private volatileServerManager serverManager
 
 
 
@@ -1590,7 +1590,7 @@ implements 
 
 assignmentManager
-privateAssignmentManager assignmentManager
+privateAssignmentManager assignmentManager
 
 
 
@@ -1599,7 +1599,7 @@ implements 
 
 replicationManager
-privateReplicationManager replicationManager
+privateReplicationManager replicationManager
 
 
 
@@ -1608,7 +1608,7 @@ implements 
 
 rsFatals
-MemoryBoundedLogMessageBuffer rsFatals
+MemoryBoundedLogMessageBuffer rsFatals
 
 
 
@@ -1617,7 +1617,7 @@ implements 
 
 activeMaster
-private volatileboolean activeMaster
+private volatileboolean activeMaster
 
 
 
@@ -1626,7 +1626,7 @@ implements 
 
 initialized
-private finalProcedureEvent initialized
+private finalProcedureEvent initialized
 
 
 
@@ -1635,7 +1635,7 @@ implements 
 
 serviceStarted
-volatileboolean serviceStarted
+volatileboolean serviceStarted
 
 
 
@@ -1644,7 +1644,7 @@ implements 
 
 serverCrashProcessingEnabled
-private finalProcedureEvent serverCrashProcessingEnabled
+private finalProcedureEvent serverCrashProcessingEnabled
 
 
 
@@ -1653,7 +1653,7 @@ implements 
 
 maxBlancingTime
-private finalint maxBlancingTime
+private 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
index c4d3f6e..bec1bbc 100644
--- a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
@@ -191,8 +191,8 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection
 org.apache.hadoop.hbase.security.AuthMethod
+org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection
 org.apache.hadoop.hbase.security.SaslStatus
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
index cd08056..570c4fa 100644
--- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
@@ -198,9 +198,9 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
+org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType
 org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType
 org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage
-org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html 
b/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
index 43c904b..58a529e 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html
@@ -320,7 +320,7 @@ extends Admin java 
class.
  Replication to secondary region replicas works over standard inter-cluster 
replication.·
 
 See Also:

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
index bc6e3ef..9aba27a 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
@@ -536,14 +536,14 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.util.ChecksumType
-org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
-org.apache.hadoop.hbase.util.PrettyPrinter.Unit
 org.apache.hadoop.hbase.util.Order
-org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE
-org.apache.hadoop.hbase.util.PoolMap.PoolType
+org.apache.hadoop.hbase.util.ChecksumType
 org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer
 (implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
+org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE
 org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer
 (implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
+org.apache.hadoop.hbase.util.PrettyPrinter.Unit
+org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
+org.apache.hadoop.hbase.util.PoolMap.PoolType
 
 
 


[27/51] [partial] hbase-site git commit: Published site at .

2017-12-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html 
b/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
index d7f2c1a..9b468f8 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
@@ -36,2607 +36,2606 @@
 028import java.util.TreeMap;
 029import java.util.concurrent.Future;
 030import java.util.regex.Pattern;
-031
-032import 
org.apache.hadoop.conf.Configuration;
-033import 
org.apache.hadoop.hbase.Abortable;
-034import 
org.apache.hadoop.hbase.CacheEvictionStats;
+031import 
org.apache.hadoop.conf.Configuration;
+032import 
org.apache.hadoop.hbase.Abortable;
+033import 
org.apache.hadoop.hbase.CacheEvictionStats;
+034import 
org.apache.hadoop.hbase.ClusterMetrics.Option;
 035import 
org.apache.hadoop.hbase.ClusterStatus;
-036import 
org.apache.hadoop.hbase.ClusterStatus.Option;
-037import 
org.apache.hadoop.hbase.HRegionInfo;
-038import 
org.apache.hadoop.hbase.HTableDescriptor;
-039import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-040import 
org.apache.hadoop.hbase.NamespaceNotFoundException;
-041import 
org.apache.hadoop.hbase.RegionLoad;
-042import 
org.apache.hadoop.hbase.ServerName;
-043import 
org.apache.hadoop.hbase.TableExistsException;
-044import 
org.apache.hadoop.hbase.TableName;
-045import 
org.apache.hadoop.hbase.TableNotFoundException;
-046import 
org.apache.yetus.audience.InterfaceAudience;
-047import 
org.apache.hadoop.hbase.client.replication.TableCFs;
-048import 
org.apache.hadoop.hbase.client.security.SecurityCapability;
-049import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
-050import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-051import 
org.apache.hadoop.hbase.quotas.QuotaRetriever;
-052import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-053import 
org.apache.hadoop.hbase.regionserver.wal.FailedLogCloseException;
-054import 
org.apache.hadoop.hbase.replication.ReplicationException;
-055import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-056import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-057import 
org.apache.hadoop.hbase.snapshot.HBaseSnapshotException;
-058import 
org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
-059import 
org.apache.hadoop.hbase.snapshot.SnapshotCreationException;
-060import 
org.apache.hadoop.hbase.snapshot.UnknownSnapshotException;
-061import 
org.apache.hadoop.hbase.util.Bytes;
-062import 
org.apache.hadoop.hbase.util.Pair;
-063
-064/**
-065 * The administrative API for HBase. 
Obtain an instance from {@link Connection#getAdmin()} and
-066 * call {@link #close()} when done.
-067 * pAdmin can be used to create, 
drop, list, enable and disable and otherwise modify tables,
-068 * as well as perform other 
administrative operations.
-069 *
-070 * @see ConnectionFactory
-071 * @see Connection
-072 * @see Table
-073 * @since 0.99.0
-074 */
-075@InterfaceAudience.Public
-076public interface Admin extends Abortable, 
Closeable {
-077  int getOperationTimeout();
-078
-079  @Override
-080  void abort(String why, Throwable e);
-081
-082  @Override
-083  boolean isAborted();
-084
-085  /**
-086   * @return Connection used by this 
object.
-087   */
-088  Connection getConnection();
-089
-090  /**
-091   * @param tableName Table to check.
-092   * @return 
codetrue/code if table exists already.
-093   * @throws IOException
-094   */
-095  boolean tableExists(TableName 
tableName) throws IOException;
-096
-097  /**
-098   * List all the userspace tables.
-099   *
-100   * @return an array of read-only 
HTableDescriptors
-101   * @throws IOException if a remote or 
network exception occurs
-102   * @deprecated since 2.0 version and 
will be removed in 3.0 version.
-103   * Use {@link 
#listTableDescriptors()}.
-104   * @see #listTableDescriptors()
-105   */
-106  @Deprecated
-107  HTableDescriptor[] listTables() throws 
IOException;
-108
-109  /**
-110   * List all the userspace tables.
-111   *
-112   * @return a list of TableDescriptors
-113   * @throws IOException if a remote or 
network exception occurs
-114   */
-115  ListTableDescriptor 
listTableDescriptors() throws IOException;
-116
-117  /**
-118   * List all the userspace tables that 
match the given pattern.
-119   *
-120   * @param pattern The compiled regular 
expression to match against
-121   * @return an array of read-only 
HTableDescriptors
-122   * @throws IOException if a remote or 
network exception occurs
-123   * @see #listTables()
-124   * @deprecated since 2.0 version and 
will be removed in 3.0 version.
-125   * Use {@link 
#listTableDescriptors(java.util.regex.Pattern)}.
-126   * @see 
#listTableDescriptors(Pattern)
-127   */
-128  @Deprecated
-129  HTableDescriptor[] listTables(Pattern 
pattern) throws 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html 
b/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html
index 6c97088..f999cbd 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html
@@ -42,679 +42,678 @@
 034
 035import 
org.apache.commons.cli.CommandLine;
 036import org.apache.commons.cli.Option;
-037import org.apache.commons.logging.Log;
-038import 
org.apache.commons.logging.LogFactory;
-039import 
org.apache.hadoop.conf.Configuration;
-040import org.apache.hadoop.fs.FileStatus;
-041import org.apache.hadoop.fs.FileSystem;
-042import org.apache.hadoop.fs.Path;
-043import 
org.apache.hadoop.hbase.TableName;
-044import 
org.apache.hadoop.hbase.client.RegionInfo;
-045import 
org.apache.hadoop.hbase.client.SnapshotDescription;
-046import 
org.apache.hadoop.hbase.io.HFileLink;
-047import 
org.apache.hadoop.hbase.io.WALLink;
-048import 
org.apache.hadoop.hbase.util.AbstractHBaseTool;
-049import 
org.apache.hadoop.hbase.util.FSUtils;
-050import 
org.apache.hadoop.util.StringUtils;
-051import 
org.apache.yetus.audience.InterfaceAudience;
-052
-053import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-054import 
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos;
-055import 
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
-056
-057/**
-058 * Tool for dumping snapshot 
information.
-059 * ol
-060 * li Table Descriptor
-061 * li Snapshot creation time, 
type, format version, ...
-062 * li List of hfiles and wals
-063 * li Stats about hfiles and logs 
sizes, percentage of shared with the source table, ...
-064 * /ol
-065 */
-066@InterfaceAudience.Public
-067public final class SnapshotInfo extends 
AbstractHBaseTool {
-068  private static final Log LOG = 
LogFactory.getLog(SnapshotInfo.class);
-069
-070  static final class Options {
-071static final Option SNAPSHOT = new 
Option(null, "snapshot", true, "Snapshot to examine.");
-072static final Option REMOTE_DIR = new 
Option(null, "remote-dir", true,
-073"Root directory that contains the 
snapshots.");
-074static final Option LIST_SNAPSHOTS = 
new Option(null, "list-snapshots", false,
-075"List all the available snapshots 
and exit.");
-076static final Option FILES = new 
Option(null, "files", false, "Files and logs list.");
-077static final Option STATS = new 
Option(null, "stats", false, "Files and logs stats.");
-078static final Option SCHEMA = new 
Option(null, "schema", false,
-079"Describe the snapshotted 
table.");
-080static final Option SIZE_IN_BYTES = 
new Option(null, "size-in-bytes", false,
-081"Print the size of the files in 
bytes.");
-082  }
-083
-084  /**
-085   * Statistics about the snapshot
-086   * ol
-087   * li How many store files and 
logs are in the archive
-088   * li How many store files and 
logs are shared with the table
-089   * li Total store files and 
logs size and shared amount
-090   * /ol
-091   */
-092  public static class SnapshotStats {
-093/** Information about the file 
referenced by the snapshot */
-094static class FileInfo {
-095  private final boolean corrupted;
-096  private final boolean inArchive;
-097  private final long size;
-098
-099  FileInfo(final boolean inArchive, 
final long size, final boolean corrupted) {
-100this.corrupted = corrupted;
-101this.inArchive = inArchive;
-102this.size = size;
-103  }
-104
-105  /** @return true if the file is in 
the archive */
-106  public boolean inArchive() {
-107return this.inArchive;
-108  }
-109
-110  /** @return true if the file is 
corrupted */
-111  public boolean isCorrupted() {
-112return this.corrupted;
-113  }
-114
-115  /** @return true if the file is 
missing */
-116  public boolean isMissing() {
-117return this.size  0;
-118  }
-119
-120  /** @return the file size */
-121  public long getSize() {
-122return this.size;
-123  }
-124
-125  String getStateToString() {
-126if (isCorrupted()) return 
"CORRUPTED";
-127if (isMissing()) return "NOT 
FOUND";
-128if (inArchive()) return 
"archive";
-129return null;
-130  }
-131}
-132
-133private AtomicInteger 
hfilesArchiveCount = new AtomicInteger();
-134private AtomicInteger hfilesCorrupted 
= new AtomicInteger();
-135private AtomicInteger hfilesMissing = 
new AtomicInteger();
-136private AtomicInteger hfilesCount = 
new AtomicInteger();
-137private AtomicInteger hfilesMobCount 
= new AtomicInteger();
-138private AtomicInteger logsMissing = 
new 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d0f1a9f6/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
index e354b98..d875570 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
@@ -129,117 +129,116 @@
 121   */
 122  @Override
 123  public void execute() throws 
IOException {
-124try (Admin admin = conn.getAdmin();) 
{
-125
-126  // Begin BACKUP
-127  beginBackup(backupManager, 
backupInfo);
-128  String savedStartCode = null;
-129  boolean firstBackup = false;
-130  // do snapshot for full table 
backup
-131
-132  savedStartCode = 
backupManager.readBackupStartCode();
-133  firstBackup = savedStartCode == 
null || Long.parseLong(savedStartCode) == 0L;
-134  if (firstBackup) {
-135// This is our first backup. 
Let's put some marker to system table so that we can hold the logs
-136// while we do the backup.
-137
backupManager.writeBackupStartCode(0L);
-138  }
-139  // We roll log here before we do 
the snapshot. It is possible there is duplicate data
-140  // in the log that is already in 
the snapshot. But if we do it after the snapshot, we
-141  // could have data loss.
-142  // A better approach is to do the 
roll log on each RS in the same global procedure as
-143  // the snapshot.
-144  LOG.info("Execute roll log 
procedure for full backup ...");
-145
-146  MapString, String props = 
new HashMapString, String();
-147  props.put("backupRoot", 
backupInfo.getBackupRootDir());
-148  
admin.execProcedure(LogRollMasterProcedureManager.ROLLLOG_PROCEDURE_SIGNATURE,
-149
LogRollMasterProcedureManager.ROLLLOG_PROCEDURE_NAME, props);
-150
-151  newTimestamps = 
backupManager.readRegionServerLastLogRollResult();
-152  if (firstBackup) {
-153// Updates registered log files
-154// We record ALL old WAL files as 
registered, because
-155// this is a first full backup in 
the system and these
-156// files are not needed for next 
incremental backup
-157ListString logFiles = 
BackupUtils.getWALFilesOlderThan(conf, newTimestamps);
-158
backupManager.recordWALFiles(logFiles);
-159  }
-160
-161  // SNAPSHOT_TABLES:
-162  
backupInfo.setPhase(BackupPhase.SNAPSHOT);
-163  for (TableName tableName : 
tableList) {
-164String snapshotName =
-165"snapshot_" + 
Long.toString(EnvironmentEdgeManager.currentTime()) + "_"
-166+ 
tableName.getNamespaceAsString() + "_" + tableName.getQualifierAsString();
-167
-168snapshotTable(admin, tableName, 
snapshotName);
-169
backupInfo.setSnapshotName(tableName, snapshotName);
-170  }
-171
-172  // SNAPSHOT_COPY:
-173  // do snapshot copy
-174  LOG.debug("snapshot copy for " + 
backupId);
-175  snapshotCopy(backupInfo);
-176  // Updates incremental backup table 
set
-177  
backupManager.addIncrementalBackupTableSet(backupInfo.getTables());
-178
-179  // BACKUP_COMPLETE:
-180  // set overall backup status: 
complete. Here we make sure to complete the backup.
-181  // After this checkpoint, even if 
entering cancel process, will let the backup finished
-182  
backupInfo.setState(BackupState.COMPLETE);
-183  // The table list in backupInfo is 
good for both full backup and incremental backup.
-184  // For incremental backup, it 
contains the incremental backup table set.
-185  
backupManager.writeRegionServerLogTimestamp(backupInfo.getTables(), 
newTimestamps);
-186
-187  HashMapTableName, 
HashMapString, Long newTableSetTimestampMap =
-188  
backupManager.readLogTimestampMap();
-189
-190  Long newStartCode =
-191  
BackupUtils.getMinValue(BackupUtils
-192  
.getRSLogTimestampMins(newTableSetTimestampMap));
-193  
backupManager.writeBackupStartCode(newStartCode);
-194
-195  // backup complete
-196  completeBackup(conn, backupInfo, 
backupManager, BackupType.FULL, conf);
-197} catch (Exception e) {
-198  failBackup(conn, backupInfo, 
backupManager, e, "Unexpected BackupException : ",
-199BackupType.FULL, conf);
-200  throw new IOException(e);
-201}
-202
-203  }
+124try (Admin admin = conn.getAdmin()) 
{
+125  // Begin BACKUP
+126  beginBackup(backupManager, 
backupInfo);
+127  String savedStartCode = null;
+128  boolean firstBackup = false;
+129  // do snapshot for full table 
backup
+130
+131  savedStartCode = 
backupManager.readBackupStartCode();
+132  firstBackup = savedStartCode == 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-16 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html 
b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
index 4dedbc2..ffeadbf 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
@@ -902,350 +902,351 @@
 894return allowPartialResults;
 895  }
 896
-897  public Scan 
setLoadColumnFamiliesOnDemand(boolean value) {
-898return (Scan) 
super.setLoadColumnFamiliesOnDemand(value);
-899  }
-900
-901  /**
-902   * Compile the table and column family 
(i.e. schema) information
-903   * into a String. Useful for parsing 
and aggregation by debugging,
-904   * logging, and administration tools.
-905   * @return Map
-906   */
-907  @Override
-908  public MapString, Object 
getFingerprint() {
-909MapString, Object map = new 
HashMap();
-910ListString families = new 
ArrayList();
-911if(this.familyMap.isEmpty()) {
-912  map.put("families", "ALL");
-913  return map;
-914} else {
-915  map.put("families", families);
-916}
-917for (Map.Entrybyte [], 
NavigableSetbyte[] entry :
-918this.familyMap.entrySet()) {
-919  
families.add(Bytes.toStringBinary(entry.getKey()));
-920}
-921return map;
-922  }
-923
-924  /**
-925   * Compile the details beyond the scope 
of getFingerprint (row, columns,
-926   * timestamps, etc.) into a Map along 
with the fingerprinted information.
-927   * Useful for debugging, logging, and 
administration tools.
-928   * @param maxCols a limit on the number 
of columns output prior to truncation
-929   * @return Map
-930   */
-931  @Override
-932  public MapString, Object 
toMap(int maxCols) {
-933// start with the fingerpring map and 
build on top of it
-934MapString, Object map = 
getFingerprint();
-935// map from families to column list 
replaces fingerprint's list of families
-936MapString, ListString 
familyColumns = new HashMap();
-937map.put("families", familyColumns);
-938// add scalar information first
-939map.put("startRow", 
Bytes.toStringBinary(this.startRow));
-940map.put("stopRow", 
Bytes.toStringBinary(this.stopRow));
-941map.put("maxVersions", 
this.maxVersions);
-942map.put("batch", this.batch);
-943map.put("caching", this.caching);
-944map.put("maxResultSize", 
this.maxResultSize);
-945map.put("cacheBlocks", 
this.cacheBlocks);
-946map.put("loadColumnFamiliesOnDemand", 
this.loadColumnFamiliesOnDemand);
-947ListLong timeRange = new 
ArrayList(2);
-948timeRange.add(this.tr.getMin());
-949timeRange.add(this.tr.getMax());
-950map.put("timeRange", timeRange);
-951int colCount = 0;
-952// iterate through affected families 
and list out up to maxCols columns
-953for (Map.Entrybyte [], 
NavigableSetbyte[] entry :
-954  this.familyMap.entrySet()) {
-955  ListString columns = new 
ArrayList();
-956  
familyColumns.put(Bytes.toStringBinary(entry.getKey()), columns);
-957  if(entry.getValue() == null) {
-958colCount++;
-959--maxCols;
-960columns.add("ALL");
-961  } else {
-962colCount += 
entry.getValue().size();
-963if (maxCols = 0) {
-964  continue;
-965}
-966for (byte [] column : 
entry.getValue()) {
-967  if (--maxCols = 0) {
-968continue;
-969  }
-970  
columns.add(Bytes.toStringBinary(column));
-971}
-972  }
-973}
-974map.put("totalColumns", colCount);
-975if (this.filter != null) {
-976  map.put("filter", 
this.filter.toString());
-977}
-978// add the id if set
-979if (getId() != null) {
-980  map.put("id", getId());
-981}
-982return map;
-983  }
-984
-985  /**
-986   * Enable/disable "raw" mode for this 
scan.
-987   * If "raw" is enabled the scan will 
return all
-988   * delete marker and deleted rows that 
have not
-989   * been collected, yet.
-990   * This is mostly useful for Scan on 
column families
-991   * that have KEEP_DELETED_ROWS 
enabled.
-992   * It is an error to specify any column 
when "raw" is set.
-993   * @param raw True/False to 
enable/disable "raw" mode.
-994   */
-995  public Scan setRaw(boolean raw) {
-996setAttribute(RAW_ATTR, 
Bytes.toBytes(raw));
-997return this;
-998  }
-999
-1000  /**
-1001   * @return True if this Scan is in 
"raw" mode.
-1002   */
-1003  public boolean isRaw() {
-1004byte[] attr = 
getAttribute(RAW_ATTR);
-1005return attr == null ? false : 
Bytes.toBoolean(attr);
-1006  }
-1007
-1008  /**
-1009   * Set whether this scan is a small 
scan
-1010   * p
-1011   * Small scan should use pread and big 
scan can use seek + read seek + read is fast but can cause
-1012   * two problem (1) resource contention 
(2) cause 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
 
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
index 54a1767..ff18b00 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public static class ThriftServerRunner.HBaseHandler
+public static class ThriftServerRunner.HBaseHandler
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 The HBaseHandler is a glue object that connects Thrift RPC 
calls to the
@@ -721,7 +721,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 conf
-protectedorg.apache.hadoop.conf.Configuration conf
+protectedorg.apache.hadoop.conf.Configuration conf
 
 
 
@@ -730,7 +730,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 LOG
-protected static finalorg.apache.commons.logging.Log LOG
+protected static finalorg.apache.commons.logging.Log LOG
 
 
 
@@ -739,7 +739,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 nextScannerId
-protectedint nextScannerId
+protectedint nextScannerId
 
 
 
@@ -748,7 +748,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 scannerMap
-protectedhttp://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true;
 title="class or interface in java.util">HashMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer,ThriftServerRunner.ResultScannerWrapper 
scannerMap
+protectedhttp://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true;
 title="class or interface in java.util">HashMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer,ThriftServerRunner.ResultScannerWrapper 
scannerMap
 
 
 
@@ -757,7 +757,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 metrics
-privateThriftMetrics metrics
+privateThriftMetrics metrics
 
 
 
@@ -766,7 +766,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 connectionCache
-private finalConnectionCache connectionCache
+private finalConnectionCache connectionCache
 
 
 
@@ -775,7 +775,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 coalescer
-IncrementCoalescer coalescer
+IncrementCoalescer coalescer
 
 
 
@@ -784,7 +784,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 CLEANUP_INTERVAL
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String CLEANUP_INTERVAL
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String CLEANUP_INTERVAL
 
 See Also:
 Constant
 Field Values
@@ -797,7 +797,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 MAX_IDLETIME
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MAX_IDLETIME
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String MAX_IDLETIME
 
 See Also:
 Constant
 Field Values
@@ -818,7 +818,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 HBaseHandler
-protectedHBaseHandler(org.apache.hadoop.conf.Configurationc,
+protectedHBaseHandler(org.apache.hadoop.conf.Configurationc,
UserProvideruserProvider)
 throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
@@ -841,7 +841,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 getAllColumns
-byte[][]getAllColumns(Tabletable)
+byte[][]getAllColumns(Tabletable)
 throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 Returns a list of all the column families for a given 
Table.
 
@@ -858,7 +858,7 @@ implements 
org.apache.hadoop.hbase.thrift.generated.Hbase.Iface
 
 
 getTable
-publicTablegetTable(byte[]tableName)
+publicTablegetTable(byte[]tableName)
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
index 7c59e27..c904c56 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html
@@ -119,4048 +119,4054 @@
 111import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
 112import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest;
 113import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearRegionBlockCacheRequest;
-114import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
-115import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
-116import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
-117import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
-118import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest;
-119import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse;
-120import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest;
-121import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest;
-122import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-123import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest;
-124import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse;
-125import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
-126import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
-127import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
-128import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-129import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos;
-130import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
-131import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
-132import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-133import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-134import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-135import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest;
-136import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-137import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-138import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest;
-139import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse;
-140import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
-141import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
-142import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
-143import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse;
-144import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest;
-145import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest;
-146import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse;
-147import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
-148import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
-149import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
-150import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
-151import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest;
-152import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse;
-153import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest;
-154import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest;
-155import 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-13 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/devapidocs/org/apache/hadoop/hbase/HConstants.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/HConstants.html 
b/devapidocs/org/apache/hadoop/hbase/HConstants.html
index 8bf6d22..2340381 100644
--- a/devapidocs/org/apache/hadoop/hbase/HConstants.html
+++ b/devapidocs/org/apache/hadoop/hbase/HConstants.html
@@ -623,279 +623,283 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
-DEFAULT_TEMPORARY_HDFS_DIRECTORY
+DEFAULT_STATUS_MULTICAST_PUBLISHER_BIND_ADDRESS
 
 
+static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+DEFAULT_TEMPORARY_HDFS_DIRECTORY
+
+
 static int
 DEFAULT_THREAD_WAKE_FREQUENCY
 Default value for thread wake frequency
 
 
-
+
 static boolean
 DEFAULT_USE_META_REPLICAS
 
-
+
 static int
 DEFAULT_VERSION_FILE_WRITE_ATTEMPTS
 Parameter name for how often we should try to write a 
version file, before failing
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 DEFAULT_WAL_STORAGE_POLICY
 
-
+
 static int
 DEFAULT_ZK_SESSION_TIMEOUT
 Default value for ZooKeeper session timeout
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 DEFAULT_ZOOKEEPER_ZNODE_PARENT
 
-
+
 static int
 DEFAULT_ZOOKEPER_CLIENT_PORT
 Default client port that the zookeeper listens on
 
 
-
+
 static int
 DEFAULT_ZOOKEPER_MAX_CLIENT_CNXNS
 Default limit on concurrent client-side zookeeper 
connections
 
 
-
+
 static int
 DELIMITER
 delimiter used between portions of a region name
 
 
-
+
 static byte[]
 EMPTY_BYTE_ARRAY
 An empty instance.
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
 title="class or interface in java.nio">ByteBuffer
 EMPTY_BYTE_BUFFER
 
-
+
 static byte[]
 EMPTY_END_ROW
 Last row in a table.
 
 
-
+
 static byte[]
 EMPTY_START_ROW
 Used by scanners, etc when they want to start at the 
beginning of a region
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 ENABLE_CLIENT_BACKPRESSURE
 Config key for if the server should send backpressure and 
if the client should listen to
  that backpressure from the server
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 ENABLE_DATA_FILE_UMASK
 Enable file permission modification from standard 
hbase
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 ENABLE_WAL_COMPRESSION
 Configuration name of WAL Compression
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 ENABLE_WAL_ENCRYPTION
 Configuration key for enabling WAL encryption, a 
boolean
 
 
-
+
 static TableName
 ENSEMBLE_TABLE_NAME
 The name of the ensemble table
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 FILE_SYSTEM_VERSION
 Current version of file system.
 
 
-
+
 static int
 FOREVER
 Unlimited time-to-live.
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 HBASE_BALANCER_MAX_BALANCING
 Config for the max balancing time
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 HBASE_BALANCER_PERIOD
 Config for the balancer period
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 HBASE_CANARY_READ_RAW_SCAN_KEY
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 HBASE_CANARY_WRITE_DATA_TTL_KEY
 Canary config keys
 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 HBASE_CANARY_WRITE_PERSERVER_REGIONS_LOWERLIMIT_KEY
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 HBASE_CANARY_WRITE_PERSERVER_REGIONS_UPPERLIMIT_KEY
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html 
b/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
index ed18c40..eb9af9e 100644
--- a/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
+++ b/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":41,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":41,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":42,"i40":42,"i41":41,"i42":42,"i43":42,"i44":42,"i45":41,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42,"i67":42,"i68":42,"i69":42,"i70":42,"i71":42,"i72":42,"i73":42,"i74":42,"i75":42,"i76":42,"i77":42,"i78":42,"i79":42,"i80":42,"i81":42,"i82":42,"i83":42};
+var methods = 
{"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":41,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":41,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":42,"i40":41,"i41":42,"i42":42,"i43":42,"i44":41,"i45":42,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42,"i67":42,"i68":42,"i69":42,"i70":42,"i71":42,"i72":42,"i73":42,"i74":42,"i75":42,"i76":42,"i77":42,"i78":42,"i79":42,"i80":42,"i81":42,"i82":42};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -792,41 +792,35 @@ implements 
 boolean
-isCacheDataInL1()
-Deprecated.
-
-
-
-boolean
 isCacheDataOnWrite()
 Deprecated.
 
 
-
+
 boolean
 isCacheIndexesOnWrite()
 Deprecated.
 
 
-
+
 boolean
 isCompressTags()
 Deprecated.
 
 
-
+
 boolean
 isEvictBlocksOnClose()
 Deprecated.
 
 
-
+
 boolean
 isInMemory()
 Deprecated.
 
 
-
+
 static byte[]
 isLegalFamilyName(byte[]b)
 Deprecated.
@@ -834,109 +828,113 @@ implements 
+
 boolean
 isMobEnabled()
 Deprecated.
 Gets whether the mob is enabled for the family.
 
 
-
+
 boolean
 isNewVersionBehavior()
 Deprecated.
 By default, HBase only consider timestamp in versions.
 
 
-
+
 boolean
 isPrefetchBlocksOnOpen()
 Deprecated.
 
 
-
+
 static HColumnDescriptor
 parseFrom(byte[]bytes)
 Deprecated.
 
 
-
+
 void
 remove(byte[]key)
 Deprecated.
 
 
-
+
 void
 removeConfiguration(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringkey)
 Deprecated.
 Remove a configuration setting represented by the key.
 
 
-
+
 HColumnDescriptor
 setBlockCacheEnabled(booleanvalue)
 Deprecated.
 
 
-
+
 HColumnDescriptor
 setBlocksize(intvalue)
 Deprecated.
 
 
-
+
 HColumnDescriptor
 setBloomFilterType(BloomTypevalue)
 Deprecated.
 
 
-
+
 HColumnDescriptor
 setCacheBloomsOnWrite(booleanvalue)
 Deprecated.
 
 
-
+
 HColumnDescriptor
 setCacheDataInL1(booleanvalue)
-Deprecated.
-
+Deprecated.
+Since 2.0 and will be 
removed in 3.0 with out any replacement. Caching data in on
+ heap Cache, when there are both on heap LRU Cache and Bucket 
Cache will no longer
+ be supported from 2.0.
+
+
 
-
+
 HColumnDescriptor
 setCacheDataOnWrite(booleanvalue)
 Deprecated.
 
 
-
+
 HColumnDescriptor
 setCacheIndexesOnWrite(booleanvalue)
 Deprecated.
 
 
-
+
 HColumnDescriptor
 setCompactionCompressionType(Compression.Algorithmvalue)
 Deprecated.
 Compression types supported in hbase.
 
 
-
+
 HColumnDescriptor
 setCompressionType(Compression.Algorithmvalue)
 Deprecated.
 Compression types supported in hbase.
 
 
-
+
 HColumnDescriptor
 setCompressTags(booleanvalue)
 Deprecated.
 Set whether the tags should be compressed along with 
DataBlockEncoding.
 
 
-
+
 HColumnDescriptor
 setConfiguration(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringkey,
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringvalue)
@@ -944,143 +942,143 @@ implements Setter for storing a configuration setting.
 
 
-
+
 HColumnDescriptor
 setDataBlockEncoding(DataBlockEncodingvalue)
 Deprecated.
 Set data block encoding algorithm used 

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-06 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html
index 00db109..9c710d6 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html
@@ -36,32 +36,32 @@
 028import java.util.NavigableMap;
 029import java.util.TreeMap;
 030import java.util.UUID;
-031
-032import org.apache.hadoop.hbase.Cell;
-033import 
org.apache.hadoop.hbase.CellScannable;
-034import 
org.apache.hadoop.hbase.CellScanner;
-035import 
org.apache.hadoop.hbase.CellUtil;
-036import 
org.apache.hadoop.hbase.HConstants;
-037import 
org.apache.hadoop.hbase.KeyValue;
-038import 
org.apache.hadoop.hbase.PrivateCellUtil;
-039import org.apache.hadoop.hbase.Tag;
-040import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-041import 
org.apache.hadoop.hbase.io.HeapSize;
-042import 
org.apache.hadoop.hbase.protobuf.ProtobufUtil;
-043import 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos;
-044import 
org.apache.hadoop.hbase.security.access.AccessControlConstants;
-045import 
org.apache.hadoop.hbase.security.access.AccessControlUtil;
-046import 
org.apache.hadoop.hbase.security.access.Permission;
-047import 
org.apache.hadoop.hbase.security.visibility.CellVisibility;
-048import 
org.apache.hadoop.hbase.security.visibility.VisibilityConstants;
-049import 
org.apache.hadoop.hbase.shaded.com.google.common.collect.ArrayListMultimap;
-050import 
org.apache.hadoop.hbase.shaded.com.google.common.collect.ListMultimap;
-051import 
org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataInput;
-052import 
org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataOutput;
-053import 
org.apache.hadoop.hbase.shaded.com.google.common.io.ByteStreams;
-054import 
org.apache.hadoop.hbase.util.Bytes;
-055import 
org.apache.hadoop.hbase.util.ClassSize;
-056import 
org.apache.yetus.audience.InterfaceAudience;
+031import org.apache.hadoop.hbase.Cell;
+032import 
org.apache.hadoop.hbase.CellScannable;
+033import 
org.apache.hadoop.hbase.CellScanner;
+034import 
org.apache.hadoop.hbase.CellUtil;
+035import 
org.apache.hadoop.hbase.HConstants;
+036import 
org.apache.hadoop.hbase.KeyValue;
+037import 
org.apache.hadoop.hbase.PrivateCellUtil;
+038import org.apache.hadoop.hbase.Tag;
+039import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
+040import 
org.apache.hadoop.hbase.io.HeapSize;
+041import 
org.apache.hadoop.hbase.protobuf.ProtobufUtil;
+042import 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos;
+043import 
org.apache.hadoop.hbase.security.access.AccessControlConstants;
+044import 
org.apache.hadoop.hbase.security.access.AccessControlUtil;
+045import 
org.apache.hadoop.hbase.security.access.Permission;
+046import 
org.apache.hadoop.hbase.security.visibility.CellVisibility;
+047import 
org.apache.hadoop.hbase.security.visibility.VisibilityConstants;
+048import 
org.apache.hadoop.hbase.util.Bytes;
+049import 
org.apache.hadoop.hbase.util.ClassSize;
+050import 
org.apache.yetus.audience.InterfaceAudience;
+051
+052import 
org.apache.hadoop.hbase.shaded.com.google.common.collect.ArrayListMultimap;
+053import 
org.apache.hadoop.hbase.shaded.com.google.common.collect.ListMultimap;
+054import 
org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataInput;
+055import 
org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataOutput;
+056import 
org.apache.hadoop.hbase.shaded.com.google.common.io.ByteStreams;
 057
 058@InterfaceAudience.Public
 059public abstract class Mutation extends 
OperationWithAttributes implements Row, CellScannable,
@@ -223,360 +223,508 @@
 215if (getTTL() != Long.MAX_VALUE) {
 216  map.put("ttl", getTTL());
 217}
-218return map;
-219  }
-220
-221  private static MapString, 
Object cellToStringMap(Cell c) {
-222MapString, Object stringMap = 
new HashMap();
-223stringMap.put("qualifier", 
Bytes.toStringBinary(c.getQualifierArray(), c.getQualifierOffset(),
-224
c.getQualifierLength()));
-225stringMap.put("timestamp", 
c.getTimestamp());
-226stringMap.put("vlen", 
c.getValueLength());
-227ListTag tags = 
PrivateCellUtil.getTags(c);
-228if (tags != null) {
-229  ListString tagsString = new 
ArrayList(tags.size());
-230  for (Tag t : tags) {
-231tagsString
-232.add((t.getType()) + ":" + 
Bytes.toStringBinary(Tag.cloneValue(t)));
-233  }
-234  stringMap.put("tag", tagsString);
-235}
-236return stringMap;
-237  }
-238
-239  /**
-240   * Set the durability for this 
mutation
-241   * @param d
-242   */
-243  public Mutation 
setDurability(Durability d) {
-244this.durability = d;
-245return this;
-246  }

[27/51] [partial] hbase-site git commit: Published site at .

2017-12-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
index 3edfbef..9707b2c 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
@@ -2459,5936 +2459,5935 @@
 2451  }
 2452
 2453  for (HStore s : storesToFlush) {
-2454MemStoreSize flushableSize = 
s.getFlushableSize();
-2455
totalSizeOfFlushableStores.incMemStoreSize(flushableSize);
-2456
storeFlushCtxs.put(s.getColumnFamilyDescriptor().getName(),
-2457  
s.createFlushContext(flushOpSeqId, tracker));
-2458// for writing stores to WAL
-2459
committedFiles.put(s.getColumnFamilyDescriptor().getName(), null);
-2460
storeFlushableSize.put(s.getColumnFamilyDescriptor().getName(), 
flushableSize);
-2461  }
-2462
-2463  // write the snapshot start to 
WAL
-2464  if (wal != null  
!writestate.readOnly) {
-2465FlushDescriptor desc = 
ProtobufUtil.toFlushDescriptor(FlushAction.START_FLUSH,
-2466getRegionInfo(), 
flushOpSeqId, committedFiles);
-2467// No sync. Sync is below where 
no updates lock and we do FlushAction.COMMIT_FLUSH
-2468WALUtil.writeFlushMarker(wal, 
this.getReplicationScope(), getRegionInfo(), desc, false,
-2469mvcc);
-2470  }
-2471
-2472  // Prepare flush (take a 
snapshot)
-2473  for (StoreFlushContext flush : 
storeFlushCtxs.values()) {
-2474flush.prepare();
-2475  }
-2476} catch (IOException ex) {
-2477  doAbortFlushToWAL(wal, 
flushOpSeqId, committedFiles);
-2478  throw ex;
-2479} finally {
-2480  
this.updatesLock.writeLock().unlock();
-2481}
-2482String s = "Finished memstore 
snapshotting " + this + ", syncing WAL and waiting on mvcc, " +
-2483"flushsize=" + 
totalSizeOfFlushableStores;
-2484status.setStatus(s);
-2485doSyncOfUnflushedWALChanges(wal, 
getRegionInfo());
-2486return new 
PrepareFlushResult(storeFlushCtxs, committedFiles, storeFlushableSize, 
startTime,
-2487flushOpSeqId, flushedSeqId, 
totalSizeOfFlushableStores);
-2488  }
-2489
-2490  /**
-2491   * Utility method broken out of 
internalPrepareFlushCache so that method is smaller.
-2492   */
-2493  private void 
logFatLineOnFlush(CollectionHStore storesToFlush, long sequenceId) {
-2494if (!LOG.isInfoEnabled()) {
-2495  return;
-2496}
-2497// Log a fat line detailing what is 
being flushed.
-2498StringBuilder perCfExtras = null;
-2499if (!isAllFamilies(storesToFlush)) 
{
-2500  perCfExtras = new 
StringBuilder();
-2501  for (HStore store: storesToFlush) 
{
-2502perCfExtras.append("; 
").append(store.getColumnFamilyName());
-2503perCfExtras.append("=")
-2504
.append(StringUtils.byteDesc(store.getFlushableSize().getDataSize()));
-2505  }
-2506}
-2507LOG.info("Flushing " + + 
storesToFlush.size() + "/" + stores.size() +
-2508" column families, memstore=" + 
StringUtils.byteDesc(this.memstoreDataSize.get()) +
-2509((perCfExtras != null  
perCfExtras.length()  0)? perCfExtras.toString(): "") +
-2510((wal != null) ? "" : "; WAL is 
null, using passed sequenceid=" + sequenceId));
-2511  }
-2512
-2513  private void doAbortFlushToWAL(final 
WAL wal, final long flushOpSeqId,
-2514  final Mapbyte[], 
ListPath committedFiles) {
-2515if (wal == null) return;
-2516try {
-2517  FlushDescriptor desc = 
ProtobufUtil.toFlushDescriptor(FlushAction.ABORT_FLUSH,
-2518  getRegionInfo(), flushOpSeqId, 
committedFiles);
-2519  WALUtil.writeFlushMarker(wal, 
this.getReplicationScope(), getRegionInfo(), desc, false,
-2520  mvcc);
-2521} catch (Throwable t) {
-2522  LOG.warn("Received unexpected 
exception trying to write ABORT_FLUSH marker to WAL:" +
-2523  
StringUtils.stringifyException(t));
-2524  // ignore this since we will be 
aborting the RS with DSE.
-2525}
-2526// we have called 
wal.startCacheFlush(), now we have to abort it
-2527
wal.abortCacheFlush(this.getRegionInfo().getEncodedNameAsBytes());
-2528  }
-2529
-2530  /**
-2531   * Sync unflushed WAL changes. See 
HBASE-8208 for details
-2532   */
-2533  private static void 
doSyncOfUnflushedWALChanges(final WAL wal, final RegionInfo hri)
-2534  throws IOException {
-2535if (wal == null) {
-2536  return;
-2537}
-2538try {
-2539  wal.sync(); // ensure that flush 
marker is sync'ed
-2540} catch (IOException ioe) {
-2541  
wal.abortCacheFlush(hri.getEncodedNameAsBytes());
-2542  throw 

[27/51] [partial] hbase-site git commit: Published site at .

2017-11-30 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
index e571d5c..3803d9a 100644
--- a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html
@@ -206,12 +206,12 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.quotas.ThrottlingException.Type
 org.apache.hadoop.hbase.quotas.OperationQuota.OperationType
-org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
-org.apache.hadoop.hbase.quotas.QuotaType
 org.apache.hadoop.hbase.quotas.QuotaScope
 org.apache.hadoop.hbase.quotas.ThrottleType
+org.apache.hadoop.hbase.quotas.QuotaType
+org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
+org.apache.hadoop.hbase.quotas.ThrottlingException.Type
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
index b091edd..ca8621b 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html
@@ -400,7 +400,7 @@ extends 
 
 triggerFlushInPrimaryRegion
-voidtriggerFlushInPrimaryRegion(HRegionregion)
+voidtriggerFlushInPrimaryRegion(HRegionregion)
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException,
  http://docs.oracle.com/javase/8/docs/api/java/lang/RuntimeException.html?is-external=true;
 title="class or interface in java.lang">RuntimeException
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index c99a865..85b1f93 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -699,19 +699,19 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
-org.apache.hadoop.hbase.regionserver.BloomType
-org.apache.hadoop.hbase.regionserver.Region.Operation
 org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result
-org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
-org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
-org.apache.hadoop.hbase.regionserver.FlushType
-org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
 org.apache.hadoop.hbase.regionserver.ScanType
+org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
 org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type
+org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
+org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
+org.apache.hadoop.hbase.regionserver.Region.Operation
+org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
+org.apache.hadoop.hbase.regionserver.FlushType
 org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
 org.apache.hadoop.hbase.regionserver.MemStoreCompactionStrategy.Action
-org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
+org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
 
 
 


[27/51] [partial] hbase-site git commit: Published site at .

2017-11-29 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html
index 91bfe40..3ddf027 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html
@@ -279,7 +279,7 @@ implements 
 Thttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in 
java.util.concurrent">CompletableFutureT
 batch(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Rowactions)
-Method that does a batch call on Deletes, Gets, Puts, 
Increments and Appends.
+Method that does a batch call on Deletes, Gets, Puts, 
Increments, Appends and RowMutations.
 
 
 
@@ -1165,15 +1165,15 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in 
java.util.concurrent">CompletableFutureTbatch(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends Rowactions)
 Description copied from 
interface:AsyncTable
-Method that does a batch call on Deletes, Gets, Puts, 
Increments and Appends. The ordering of
- execution of the actions is not defined. Meaning if you do a Put and a Get in 
the same
- AsyncTable.batch(java.util.List?
 extends org.apache.hadoop.hbase.client.Row) call, you will not 
necessarily be guaranteed that the Get returns what the Put
- had put.
+Method that does a batch call on Deletes, Gets, Puts, 
Increments, Appends and RowMutations. The
+ ordering of execution of the actions is not defined. Meaning if you do a Put 
and a Get in the
+ same AsyncTable.batch(java.util.List?
 extends org.apache.hadoop.hbase.client.Row) call, you will not 
necessarily be guaranteed that the Get returns what the
+ Put had put.
 
 Specified by:
 batchin
 interfaceAsyncTableAdvancedScanResultConsumer
 Parameters:
-actions - list of Get, Put, Delete, Increment, Append 
objects
+actions - list of Get, Put, Delete, Increment, Append, and 
RowMutations objects
 Returns:
 A list of http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in 
java.util.concurrent">CompletableFutures that represent the 
result for each action.
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html 
b/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html
index 495b3bf..66c3237 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":17,"i1":6,"i2":6,"i3":17,"i4":17,"i5":17,"i6":17,"i7":17,"i8":17,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":17,"i18":17,"i19":6,"i20":17,"i21":6,"i22":17,"i23":17,"i24":17,"i25":6,"i26":6,"i27":6,"i28":6,"i29":17,"i30":17,"i31":17,"i32":17,"i33":17,"i34":17,"i35":17,"i36":17,"i37":17,"i38":17,"i39":17};
+var methods = 
{"i0":17,"i1":6,"i2":6,"i3":17,"i4":17,"i5":17,"i6":17,"i7":17,"i8":17,"i9":17,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":17,"i19":17,"i20":6,"i21":17,"i22":6,"i23":17,"i24":17,"i25":17,"i26":6,"i27":6,"i28":6,"i29":6,"i30":17,"i31":17,"i32":17,"i33":17,"i34":17,"i35":17,"i36":17,"i37":17,"i38":17,"i39":17,"i40":17};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],4:["t3","Abstract 
Methods"],16:["t5","Default Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -212,6 +212,12 @@ public interface containsRow(byte[]row)
 
 
+static RegionInfo
+createMobRegionInfo(TableNametableName)
+Creates a RegionInfo object for MOB data.
+
+
+
 static byte[]
 createRegionName(TableNametableName,
 byte[]startKey,
@@ -220,7 +226,7 @@ public interface Make a region name of passed parameters.
 
 
-
+
 static byte[]
 createRegionName(TableNametableName,
 byte[]startKey,
@@ -230,7 +236,7 @@ public interface Make a region name of passed parameters.
 
 
-
+
 static byte[]
 createRegionName(TableNametableName,
 byte[]startKey,
@@ -239,7 +245,7 

[27/51] [partial] hbase-site git commit: Published site at .

2017-11-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9722a17/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
index 860d344..9fa61e7 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
@@ -1604,71 +1604,72 @@
 1596});
 1597  }
 1598
-1599  public void postClearDeadServers() 
throws IOException {
-1600
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-1601  @Override
-1602  public void call(MasterObserver 
observer) throws IOException {
-1603
observer.postClearDeadServers(this);
-1604  }
-1605});
-1606  }
-1607
-1608  public void 
preDecommissionRegionServers(ListServerName servers, boolean offload) 
throws IOException {
-1609
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-1610  @Override
-1611  public void call(MasterObserver 
observer) throws IOException {
-1612
observer.preDecommissionRegionServers(this, servers, offload);
-1613  }
-1614});
-1615  }
-1616
-1617  public void 
postDecommissionRegionServers(ListServerName servers, boolean offload) 
throws IOException {
-1618
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-1619  @Override
-1620  public void call(MasterObserver 
observer) throws IOException {
-1621
observer.postDecommissionRegionServers(this, servers, offload);
-1622  }
-1623});
-1624  }
-1625
-1626  public void 
preListDecommissionedRegionServers() throws IOException {
-1627
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-1628  @Override
-1629  public void call(MasterObserver 
observer) throws IOException {
-1630
observer.preListDecommissionedRegionServers(this);
-1631  }
-1632});
-1633  }
-1634
-1635  public void 
postListDecommissionedRegionServers() throws IOException {
-1636
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-1637  @Override
-1638  public void call(MasterObserver 
observer) throws IOException {
-1639
observer.postListDecommissionedRegionServers(this);
-1640  }
-1641});
-1642  }
-1643
-1644  public void 
preRecommissionRegionServer(ServerName server, Listbyte[] 
encodedRegionNames)
-1645  throws IOException {
-1646
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-1647  @Override
-1648  public void call(MasterObserver 
observer) throws IOException {
-1649
observer.preRecommissionRegionServer(this, server, encodedRegionNames);
-1650  }
-1651});
-1652  }
-1653
-1654  public void 
postRecommissionRegionServer(ServerName server, Listbyte[] 
encodedRegionNames)
-1655  throws IOException {
-1656
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
-1657  @Override
-1658  public void call(MasterObserver 
observer) throws IOException {
-1659
observer.postRecommissionRegionServer(this, server, encodedRegionNames);
-1660  }
-1661});
-1662  }
-1663}
+1599  public void 
postClearDeadServers(ListServerName servers,
+1600  ListServerName 
notClearedServers) throws IOException {
+1601
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
+1602  @Override
+1603  public void call(MasterObserver 
observer) throws IOException {
+1604
observer.postClearDeadServers(this, servers, notClearedServers);
+1605  }
+1606});
+1607  }
+1608
+1609  public void 
preDecommissionRegionServers(ListServerName servers, boolean offload) 
throws IOException {
+1610
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
+1611  @Override
+1612  public void call(MasterObserver 
observer) throws IOException {
+1613
observer.preDecommissionRegionServers(this, servers, offload);
+1614  }
+1615});
+1616  }
+1617
+1618  public void 
postDecommissionRegionServers(ListServerName servers, boolean offload) 
throws IOException {
+1619
execOperation(coprocEnvironments.isEmpty() ? null : new 
MasterObserverOperation() {
+1620  @Override
+1621  public void call(MasterObserver 
observer) throws IOException {
+1622
observer.postDecommissionRegionServers(this, servers, offload);
+1623  }
+1624});
+1625  }
+1626
+1627  public void 
preListDecommissionedRegionServers() throws IOException {
+1628

  1   2   3   4   >