hbase git commit: HBASE-19187 Remove option to create on heap bucket cache.

2017-11-10 Thread anoopsamjohn
Repository: hbase
Updated Branches:
  refs/heads/branch-2 30f55f231 -> 3ee76f857


HBASE-19187 Remove option to create on heap bucket cache.


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/3ee76f85
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/3ee76f85
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/3ee76f85

Branch: refs/heads/branch-2
Commit: 3ee76f8573387097c896d0611c30504ed7e57cb0
Parents: 30f55f2
Author: anoopsamjohn 
Authored: Sat Nov 11 12:37:18 2017 +0530
Committer: anoopsamjohn 
Committed: Sat Nov 11 12:41:09 2017 +0530

--
 .../hadoop/hbase/util/ByteBufferAllocator.java  |  3 +-
 .../hadoop/hbase/util/ByteBufferArray.java  | 18 +++
 .../src/main/resources/hbase-default.xml|  7 +--
 .../hadoop/hbase/util/TestByteBufferArray.java  | 38 +-
 .../hbase/io/hfile/bucket/BucketCache.java  |  6 +--
 .../io/hfile/bucket/ByteBufferIOEngine.java | 18 ++-
 .../hbase/io/hfile/bucket/FileMmapEngine.java   | 14 ++
 .../hadoop/hbase/io/util/MemorySizeUtil.java| 44 +++-
 .../hbase/regionserver/HeapMemoryManager.java   |  8 ++-
 ...estAvoidCellReferencesIntoShippedBlocks.java |  2 +-
 .../client/TestBlockEvictionFromClient.java |  2 +-
 .../hadoop/hbase/io/hfile/TestCacheConfig.java  |  6 ---
 .../io/hfile/TestScannerFromBucketCache.java| 53 +++-
 .../hbase/io/hfile/bucket/TestBucketCache.java  |  4 +-
 .../io/hfile/bucket/TestBucketWriterThread.java |  3 +-
 .../io/hfile/bucket/TestByteBufferIOEngine.java |  4 +-
 .../regionserver/TestHeapMemoryManager.java | 52 ---
 17 files changed, 57 insertions(+), 225 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/3ee76f85/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
index 8e2ac16..2c5eac8 100644
--- 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
+++ 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
@@ -31,9 +31,8 @@ public interface ByteBufferAllocator {
   /**
* Allocates a bytebuffer
* @param size the size of the bytebuffer
-   * @param directByteBuffer indicator to create a direct bytebuffer
* @return the bytebuffer that is created
* @throws IOException exception thrown if there is an error while creating 
the ByteBuffer
*/
-  ByteBuffer allocate(long size, boolean directByteBuffer) throws IOException;
+  ByteBuffer allocate(long size) throws IOException;
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/3ee76f85/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java
index d1d5e7d..2715740 100644
--- 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java
+++ 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java
@@ -59,11 +59,10 @@ public class ByteBufferArray {
* of the array bounds for the last byte(see {@link 
ByteBufferArray#multiple}),
* we will allocate one additional buffer with capacity 0;
* @param capacity total size of the byte buffer array
-   * @param directByteBuffer true if we allocate direct buffer
* @param allocator the ByteBufferAllocator that will create the buffers
* @throws IOException throws IOException if there is an exception thrown by 
the allocator
*/
-  public ByteBufferArray(long capacity, boolean directByteBuffer, 
ByteBufferAllocator allocator)
+  public ByteBufferArray(long capacity, ByteBufferAllocator allocator)
   throws IOException {
 this.bufferSize = DEFAULT_BUFFER_SIZE;
 if (this.bufferSize > (capacity / 16))
@@ -71,13 +70,13 @@ public class ByteBufferArray {
 this.bufferCount = (int) (roundUp(capacity, bufferSize) / bufferSize);
 LOG.info("Allocating buffers total=" + StringUtils.byteDesc(capacity)
 + ", sizePerBuffer=" + StringUtils.byteDesc(bufferSize) + ", count="
-+ bufferCount + ", direct=" + directByteBuffer);
++ bufferCount);
 buffers = new ByteBuffer[bufferCount + 1];
-createBuffers(directByteBuffer, allocator);
+createBuffers(allocator);
   }
 
   @VisibleForTesting
-  void createBuffers(boolean directByteBuffer, ByteBufferAllocator allocator)
+  void 

hbase git commit: HBASE-19187 Remove option to create on heap bucket cache.

2017-11-10 Thread anoopsamjohn
Repository: hbase
Updated Branches:
  refs/heads/master 72270866c -> bff619ef7


HBASE-19187 Remove option to create on heap bucket cache.


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/bff619ef
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/bff619ef
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/bff619ef

Branch: refs/heads/master
Commit: bff619ef7b100e8b09f7f5eb0f6e289ca51de096
Parents: 7227086
Author: anoopsamjohn 
Authored: Sat Nov 11 12:37:18 2017 +0530
Committer: anoopsamjohn 
Committed: Sat Nov 11 12:37:18 2017 +0530

--
 .../hadoop/hbase/util/ByteBufferAllocator.java  |  3 +-
 .../hadoop/hbase/util/ByteBufferArray.java  | 18 +++
 .../src/main/resources/hbase-default.xml|  7 +--
 .../hadoop/hbase/util/TestByteBufferArray.java  | 38 +-
 .../hbase/io/hfile/bucket/BucketCache.java  |  6 +--
 .../io/hfile/bucket/ByteBufferIOEngine.java | 18 ++-
 .../hbase/io/hfile/bucket/FileMmapEngine.java   | 14 ++
 .../hadoop/hbase/io/util/MemorySizeUtil.java| 44 +++-
 .../hbase/regionserver/HeapMemoryManager.java   |  8 ++-
 ...estAvoidCellReferencesIntoShippedBlocks.java |  2 +-
 .../client/TestBlockEvictionFromClient.java |  2 +-
 .../hadoop/hbase/io/hfile/TestCacheConfig.java  |  6 ---
 .../io/hfile/TestScannerFromBucketCache.java| 53 +++-
 .../hbase/io/hfile/bucket/TestBucketCache.java  |  4 +-
 .../io/hfile/bucket/TestBucketWriterThread.java |  3 +-
 .../io/hfile/bucket/TestByteBufferIOEngine.java |  4 +-
 .../regionserver/TestHeapMemoryManager.java | 52 ---
 17 files changed, 57 insertions(+), 225 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/bff619ef/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
index 8e2ac16..2c5eac8 100644
--- 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
+++ 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java
@@ -31,9 +31,8 @@ public interface ByteBufferAllocator {
   /**
* Allocates a bytebuffer
* @param size the size of the bytebuffer
-   * @param directByteBuffer indicator to create a direct bytebuffer
* @return the bytebuffer that is created
* @throws IOException exception thrown if there is an error while creating 
the ByteBuffer
*/
-  ByteBuffer allocate(long size, boolean directByteBuffer) throws IOException;
+  ByteBuffer allocate(long size) throws IOException;
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bff619ef/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java
index d1d5e7d..2715740 100644
--- 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java
+++ 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java
@@ -59,11 +59,10 @@ public class ByteBufferArray {
* of the array bounds for the last byte(see {@link 
ByteBufferArray#multiple}),
* we will allocate one additional buffer with capacity 0;
* @param capacity total size of the byte buffer array
-   * @param directByteBuffer true if we allocate direct buffer
* @param allocator the ByteBufferAllocator that will create the buffers
* @throws IOException throws IOException if there is an exception thrown by 
the allocator
*/
-  public ByteBufferArray(long capacity, boolean directByteBuffer, 
ByteBufferAllocator allocator)
+  public ByteBufferArray(long capacity, ByteBufferAllocator allocator)
   throws IOException {
 this.bufferSize = DEFAULT_BUFFER_SIZE;
 if (this.bufferSize > (capacity / 16))
@@ -71,13 +70,13 @@ public class ByteBufferArray {
 this.bufferCount = (int) (roundUp(capacity, bufferSize) / bufferSize);
 LOG.info("Allocating buffers total=" + StringUtils.byteDesc(capacity)
 + ", sizePerBuffer=" + StringUtils.byteDesc(bufferSize) + ", count="
-+ bufferCount + ", direct=" + directByteBuffer);
++ bufferCount);
 buffers = new ByteBuffer[bufferCount + 1];
-createBuffers(directByteBuffer, allocator);
+createBuffers(allocator);
   }
 
   @VisibleForTesting
-  void createBuffers(boolean directByteBuffer, ByteBufferAllocator allocator)
+  void 

[2/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths

2017-11-10 Thread appy
http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
index 2a0764d..5954e94 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
@@ -22,22 +22,24 @@ import static 
org.apache.hadoop.hbase.HConstants.DEFAULT_ZOOKEEPER_ZNODE_PARENT;
 import static org.apache.hadoop.hbase.HConstants.META_REPLICAS_NUM;
 import static org.apache.hadoop.hbase.HConstants.SPLIT_LOGDIR_NAME;
 import static org.apache.hadoop.hbase.HConstants.ZOOKEEPER_ZNODE_PARENT;
-import static org.apache.hadoop.hbase.HRegionInfo.DEFAULT_REPLICA_ID;
-
-import org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap;
+import static org.apache.hadoop.hbase.client.RegionInfo.DEFAULT_REPLICA_ID;
 
 import java.util.Optional;
 import java.util.stream.IntStream;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hbase.HRegionInfo;
+import org.apache.hadoop.hbase.client.RegionInfo;
 import org.apache.yetus.audience.InterfaceAudience;
 
+import org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap;
+
 /**
  * Class that hold all the paths of znode for HBase.
  */
 @InterfaceAudience.Private
 public class ZNodePaths {
+  // TODO: Replace this with ZooKeeper constant when ZOOKEEPER-277 is resolved.
+  public static final char ZNODE_PATH_SEPARATOR = '/';
 
   public final static String META_ZNODE_PREFIX = "meta-region-server";
 
@@ -90,43 +92,35 @@ public class ZNodePaths {
 baseZNode = conf.get(ZOOKEEPER_ZNODE_PARENT, 
DEFAULT_ZOOKEEPER_ZNODE_PARENT);
 ImmutableMap.Builder builder = ImmutableMap.builder();
 metaZNodePrefix = conf.get("zookeeper.znode.metaserver", 
META_ZNODE_PREFIX);
-String defaultMetaReplicaZNode = ZKUtil.joinZNode(baseZNode, 
metaZNodePrefix);
+String defaultMetaReplicaZNode = ZNodePaths.joinZNode(baseZNode, 
metaZNodePrefix);
 builder.put(DEFAULT_REPLICA_ID, defaultMetaReplicaZNode);
 int numMetaReplicas = conf.getInt(META_REPLICAS_NUM, 
DEFAULT_META_REPLICA_NUM);
 IntStream.range(1, numMetaReplicas)
 .forEachOrdered(i -> builder.put(i, defaultMetaReplicaZNode + "-" + 
i));
 metaReplicaZNodes = builder.build();
-rsZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.rs", 
"rs"));
-drainingZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.draining.rs", "draining"));
-masterAddressZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.master", "master"));
-backupMasterAddressesZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.backup.masters", "backup-masters"));
-clusterStateZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.state", "running"));
-tableZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.tableEnableDisable", "table"));
-clusterIdZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.clusterId", "hbaseid"));
-splitLogZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.splitlog", SPLIT_LOGDIR_NAME));
-balancerZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.balancer", "balancer"));
-regionNormalizerZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.regionNormalizer", "normalizer"));
-switchZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.switch", "switch"));
-tableLockZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.tableLock", "table-lock"));
-namespaceZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.namespace", "namespace"));
-masterMaintZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.masterMaintenance", "master-maintenance"));
-replicationZNode =
-ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.replication", 
"replication"));
+rsZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.rs", "rs"));
+drainingZNode = joinZNode(baseZNode, 
conf.get("zookeeper.znode.draining.rs", "draining"));
+masterAddressZNode = joinZNode(baseZNode, 
conf.get("zookeeper.znode.master", "master"));
+backupMasterAddressesZNode =
+joinZNode(baseZNode, conf.get("zookeeper.znode.backup.masters", 
"backup-masters"));
+clusterStateZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.state", 
"running"));
+tableZNode = joinZNode(baseZNode, 
conf.get("zookeeper.znode.tableEnableDisable", "table"));
+clusterIdZNode = joinZNode(baseZNode, 
conf.get("zookeeper.znode.clusterId", "hbaseid"));
+splitLogZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.splitlog", 

[3/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths

2017-11-10 Thread appy
HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths

- Removes zookeeper connection from ClusterConnection
- Deletes class ZooKeeperKeepAliveConnection
- Removes Registry, ZooKeeperRegistry, and RegistryFactory


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/30f55f23
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/30f55f23
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/30f55f23

Branch: refs/heads/branch-2
Commit: 30f55f23166f3b0652bb2a96f8bbb1b2bcde4ace
Parents: dd19fa2
Author: zhangduo 
Authored: Fri Nov 10 22:16:45 2017 +0800
Committer: Apekshit Sharma 
Committed: Fri Nov 10 10:09:04 2017 -0800

--
 .../hadoop/hbase/client/ClusterConnection.java  |   4 +
 .../hbase/client/ConnectionImplementation.java  | 236 +-
 .../apache/hadoop/hbase/client/HBaseAdmin.java  | 326 +++
 .../apache/hadoop/hbase/client/MetaCache.java   |  27 +-
 .../apache/hadoop/hbase/client/Registry.java|  53 ---
 .../hadoop/hbase/client/RegistryFactory.java|  50 ---
 .../hadoop/hbase/client/ZKAsyncRegistry.java|  13 +-
 .../client/ZooKeeperKeepAliveConnection.java|  56 
 .../hadoop/hbase/client/ZooKeeperRegistry.java  | 129 
 .../hbase/zookeeper/MasterAddressTracker.java   |   2 +-
 .../hbase/zookeeper/RecoverableZooKeeper.java   |  70 +---
 .../hadoop/hbase/zookeeper/ZKMetadata.java  |  80 +
 .../apache/hadoop/hbase/zookeeper/ZKUtil.java   |  35 +-
 .../hadoop/hbase/zookeeper/ZNodePaths.java  |  80 +++--
 .../hadoop/hbase/zookeeper/ZkAclReset.java  |   2 +-
 .../hbase/zookeeper/ZooKeeperWatcher.java   |   2 +-
 .../hbase/client/DoNothingAsyncRegistry.java|  64 
 .../hadoop/hbase/client/TestAsyncProcess.java   |  21 +-
 .../hbase/client/TestBufferedMutator.java   |  33 +-
 .../hbase/client/TestClientNoCluster.java   |  66 ++--
 .../hbase/zookeeper/TestZooKeeperWatcher.java   |   9 +-
 .../hbase/IntegrationTestMetaReplicas.java  |   3 +-
 .../test/IntegrationTestZKAndFSPermissions.java |   3 +-
 .../replication/ReplicationPeersZKImpl.java |   3 +-
 .../ReplicationQueuesClientZKImpl.java  |   9 +-
 .../replication/ReplicationQueuesZKImpl.java|  54 +--
 .../replication/ReplicationStateZKBase.java |  17 +-
 .../hbase/rsgroup/RSGroupInfoManagerImpl.java   |  11 +-
 .../rsgroup/VerifyingRSGroupAdminClient.java|   5 +-
 .../apache/hadoop/hbase/ZKNamespaceManager.java |   5 +-
 .../org/apache/hadoop/hbase/ZNodeClearer.java   |  12 +-
 .../backup/example/HFileArchiveManager.java |   3 +-
 .../backup/example/ZKTableArchiveClient.java|   4 +-
 .../ZKSplitLogManagerCoordination.java  |  10 +-
 .../ZkSplitLogWorkerCoordination.java   |   9 +-
 .../hbase/master/ActiveMasterManager.java   |   3 +-
 .../org/apache/hadoop/hbase/master/HMaster.java |   9 +-
 .../cleaner/ReplicationZKNodeCleaner.java   |   7 +-
 .../hbase/procedure/ZKProcedureCoordinator.java |   5 +-
 .../hbase/procedure/ZKProcedureMemberRpcs.java  |  10 +-
 .../hadoop/hbase/procedure/ZKProcedureUtil.java |  19 +-
 .../hbase/regionserver/HRegionServer.java   |   3 +-
 .../security/access/ZKPermissionWatcher.java|  15 +-
 .../token/AuthenticationTokenSecretManager.java |   4 +-
 .../hbase/security/token/ZKSecretWatcher.java   |   7 +-
 .../visibility/ZKVisibilityLabelWatcher.java|   5 +-
 .../org/apache/hadoop/hbase/util/HBaseFsck.java |   3 +-
 .../hadoop/hbase/util/ZKDataMigrator.java   |   3 +-
 .../hbase/zookeeper/RegionServerTracker.java|   2 +-
 .../hbase/zookeeper/SplitOrMergeTracker.java|   4 +-
 .../hadoop/hbase/zookeeper/ZKSplitLog.java  |   4 +-
 .../apache/hadoop/hbase/GenericTestUtils.java   |   1 +
 .../hbase/client/TestMetaWithReplicas.java  |   9 +-
 .../hbase/master/TestAssignmentListener.java|   5 +-
 .../hbase/master/TestMetaShutdownHandler.java   |   3 +-
 .../hbase/master/TestTableStateManager.java |   3 +-
 .../TestCompactionInDeadRegionServer.java   |   4 +-
 .../hbase/regionserver/TestSplitLogWorker.java  |   7 +-
 .../replication/TestMasterReplication.java  |   7 +-
 .../TestReplicationStateHBaseImpl.java  |   3 +-
 .../replication/TestReplicationStateZKImpl.java |   9 +-
 .../TestReplicationTrackerZKImpl.java   |  18 +-
 .../TestReplicationSourceManager.java   |   8 +-
 .../token/TestZKSecretWatcherRefreshKeys.java   |   3 +-
 .../hadoop/hbase/zookeeper/TestZKMulti.java |  40 +--
 .../zookeeper/TestZooKeeperNodeTracker.java |   2 +-
 66 files changed, 687 insertions(+), 1044 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java

[1/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths

2017-11-10 Thread appy
Repository: hbase
Updated Branches:
  refs/heads/branch-2 dd19fa25d -> 30f55f231


http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
index 9b5a44a..5a6a414 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
@@ -24,6 +24,7 @@ import org.apache.commons.logging.LogFactory;
 import org.apache.yetus.audience.InterfaceAudience;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.zookeeper.ZKUtil;
+import org.apache.hadoop.hbase.zookeeper.ZNodePaths;
 import org.apache.hadoop.hbase.zookeeper.ZooKeeperListener;
 import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
 import org.apache.zookeeper.KeeperException;
@@ -53,8 +54,8 @@ public class ZKVisibilityLabelWatcher extends 
ZooKeeperListener {
 String labelZnodeParent = conf.get(VISIBILITY_LABEL_ZK_PATH, 
DEFAULT_VISIBILITY_LABEL_NODE);
 String userAuthsZnodeParent = conf.get(VISIBILITY_USER_AUTHS_ZK_PATH,
 DEFAULT_VISIBILITY_USER_AUTHS_NODE);
-this.labelZnode = ZKUtil.joinZNode(watcher.znodePaths.baseZNode, 
labelZnodeParent);
-this.userAuthsZnode = ZKUtil.joinZNode(watcher.znodePaths.baseZNode, 
userAuthsZnodeParent);
+this.labelZnode = ZNodePaths.joinZNode(watcher.znodePaths.baseZNode, 
labelZnodeParent);
+this.userAuthsZnode = ZNodePaths.joinZNode(watcher.znodePaths.baseZNode, 
userAuthsZnodeParent);
   }
 
   public void start() throws KeeperException {

http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
index 934a630..bf4478d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
@@ -130,6 +130,7 @@ import org.apache.hadoop.hbase.wal.WALFactory;
 import org.apache.hadoop.hbase.wal.WALSplitter;
 import org.apache.hadoop.hbase.zookeeper.MetaTableLocator;
 import org.apache.hadoop.hbase.zookeeper.ZKUtil;
+import org.apache.hadoop.hbase.zookeeper.ZNodePaths;
 import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
 import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException;
 import org.apache.hadoop.ipc.RemoteException;
@@ -703,7 +704,7 @@ public class HBaseFsck extends Configured implements 
Closeable {
*/
   private boolean setMasterInMaintenanceMode() throws IOException {
 RetryCounter retryCounter = createZNodeRetryCounterFactory.create();
-hbckEphemeralNodePath = ZKUtil.joinZNode(
+hbckEphemeralNodePath = ZNodePaths.joinZNode(
   zkw.znodePaths.masterMaintZNode,
   "hbck-" + Long.toString(EnvironmentEdgeManager.currentTime()));
 do {

http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java
index 43c3598..e07861e 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java
@@ -31,6 +31,7 @@ import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
 import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos;
 import org.apache.hadoop.hbase.zookeeper.ZKUtil;
+import org.apache.hadoop.hbase.zookeeper.ZNodePaths;
 import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
 import org.apache.zookeeper.KeeperException;
 
@@ -94,7 +95,7 @@ public class ZKDataMigrator {
   private static  ZooKeeperProtos.DeprecatedTableState.State getTableState(
   final ZooKeeperWatcher zkw, final TableName tableName)
   throws KeeperException, InterruptedException {
-String znode = ZKUtil.joinZNode(zkw.znodePaths.tableZNode, 
tableName.getNameAsString());
+String znode = ZNodePaths.joinZNode(zkw.znodePaths.tableZNode, 
tableName.getNameAsString());
 byte [] data = ZKUtil.getData(zkw, znode);
 if (data == null || data.length <= 0) return null;
 try {


[1/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths

2017-11-10 Thread appy
Repository: hbase
Updated Branches:
  refs/heads/master 31234eb86 -> 72270866c


http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
index 9b5a44a..5a6a414 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java
@@ -24,6 +24,7 @@ import org.apache.commons.logging.LogFactory;
 import org.apache.yetus.audience.InterfaceAudience;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.zookeeper.ZKUtil;
+import org.apache.hadoop.hbase.zookeeper.ZNodePaths;
 import org.apache.hadoop.hbase.zookeeper.ZooKeeperListener;
 import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
 import org.apache.zookeeper.KeeperException;
@@ -53,8 +54,8 @@ public class ZKVisibilityLabelWatcher extends 
ZooKeeperListener {
 String labelZnodeParent = conf.get(VISIBILITY_LABEL_ZK_PATH, 
DEFAULT_VISIBILITY_LABEL_NODE);
 String userAuthsZnodeParent = conf.get(VISIBILITY_USER_AUTHS_ZK_PATH,
 DEFAULT_VISIBILITY_USER_AUTHS_NODE);
-this.labelZnode = ZKUtil.joinZNode(watcher.znodePaths.baseZNode, 
labelZnodeParent);
-this.userAuthsZnode = ZKUtil.joinZNode(watcher.znodePaths.baseZNode, 
userAuthsZnodeParent);
+this.labelZnode = ZNodePaths.joinZNode(watcher.znodePaths.baseZNode, 
labelZnodeParent);
+this.userAuthsZnode = ZNodePaths.joinZNode(watcher.znodePaths.baseZNode, 
userAuthsZnodeParent);
   }
 
   public void start() throws KeeperException {

http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
index 934a630..bf4478d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java
@@ -130,6 +130,7 @@ import org.apache.hadoop.hbase.wal.WALFactory;
 import org.apache.hadoop.hbase.wal.WALSplitter;
 import org.apache.hadoop.hbase.zookeeper.MetaTableLocator;
 import org.apache.hadoop.hbase.zookeeper.ZKUtil;
+import org.apache.hadoop.hbase.zookeeper.ZNodePaths;
 import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
 import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException;
 import org.apache.hadoop.ipc.RemoteException;
@@ -703,7 +704,7 @@ public class HBaseFsck extends Configured implements 
Closeable {
*/
   private boolean setMasterInMaintenanceMode() throws IOException {
 RetryCounter retryCounter = createZNodeRetryCounterFactory.create();
-hbckEphemeralNodePath = ZKUtil.joinZNode(
+hbckEphemeralNodePath = ZNodePaths.joinZNode(
   zkw.znodePaths.masterMaintZNode,
   "hbck-" + Long.toString(EnvironmentEdgeManager.currentTime()));
 do {

http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java
index 43c3598..e07861e 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java
@@ -31,6 +31,7 @@ import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
 import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos;
 import org.apache.hadoop.hbase.zookeeper.ZKUtil;
+import org.apache.hadoop.hbase.zookeeper.ZNodePaths;
 import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
 import org.apache.zookeeper.KeeperException;
 
@@ -94,7 +95,7 @@ public class ZKDataMigrator {
   private static  ZooKeeperProtos.DeprecatedTableState.State getTableState(
   final ZooKeeperWatcher zkw, final TableName tableName)
   throws KeeperException, InterruptedException {
-String znode = ZKUtil.joinZNode(zkw.znodePaths.tableZNode, 
tableName.getNameAsString());
+String znode = ZNodePaths.joinZNode(zkw.znodePaths.tableZNode, 
tableName.getNameAsString());
 byte [] data = ZKUtil.getData(zkw, znode);
 if (data == null || data.length <= 0) return null;
 try {


[3/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths

2017-11-10 Thread appy
HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths

- Removes zookeeper connection from ClusterConnection
- Deletes class ZooKeeperKeepAliveConnection
- Removes Registry, ZooKeeperRegistry, and RegistryFactory


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/72270866
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/72270866
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/72270866

Branch: refs/heads/master
Commit: 72270866cbacbcfd10f2ad2cce33362577eeef4d
Parents: 31234eb
Author: zhangduo 
Authored: Fri Nov 10 22:16:45 2017 +0800
Committer: Apekshit Sharma 
Committed: Fri Nov 10 10:05:25 2017 -0800

--
 .../hadoop/hbase/client/ClusterConnection.java  |   4 +
 .../hbase/client/ConnectionImplementation.java  | 236 +-
 .../apache/hadoop/hbase/client/HBaseAdmin.java  | 326 +++
 .../apache/hadoop/hbase/client/MetaCache.java   |  27 +-
 .../apache/hadoop/hbase/client/Registry.java|  53 ---
 .../hadoop/hbase/client/RegistryFactory.java|  50 ---
 .../hadoop/hbase/client/ZKAsyncRegistry.java|  13 +-
 .../client/ZooKeeperKeepAliveConnection.java|  56 
 .../hadoop/hbase/client/ZooKeeperRegistry.java  | 129 
 .../hbase/zookeeper/MasterAddressTracker.java   |   2 +-
 .../hbase/zookeeper/RecoverableZooKeeper.java   |  70 +---
 .../hadoop/hbase/zookeeper/ZKMetadata.java  |  80 +
 .../apache/hadoop/hbase/zookeeper/ZKUtil.java   |  35 +-
 .../hadoop/hbase/zookeeper/ZNodePaths.java  |  80 +++--
 .../hadoop/hbase/zookeeper/ZkAclReset.java  |   2 +-
 .../hbase/zookeeper/ZooKeeperWatcher.java   |   2 +-
 .../hbase/client/DoNothingAsyncRegistry.java|  64 
 .../hadoop/hbase/client/TestAsyncProcess.java   |  21 +-
 .../hbase/client/TestBufferedMutator.java   |  33 +-
 .../hbase/client/TestClientNoCluster.java   |  66 ++--
 .../hbase/zookeeper/TestZooKeeperWatcher.java   |   9 +-
 .../hbase/IntegrationTestMetaReplicas.java  |   3 +-
 .../test/IntegrationTestZKAndFSPermissions.java |   3 +-
 .../replication/ReplicationPeersZKImpl.java |   3 +-
 .../ReplicationQueuesClientZKImpl.java  |   9 +-
 .../replication/ReplicationQueuesZKImpl.java|  54 +--
 .../replication/ReplicationStateZKBase.java |  17 +-
 .../hbase/rsgroup/RSGroupInfoManagerImpl.java   |  11 +-
 .../rsgroup/VerifyingRSGroupAdminClient.java|   5 +-
 .../apache/hadoop/hbase/ZKNamespaceManager.java |   5 +-
 .../org/apache/hadoop/hbase/ZNodeClearer.java   |  12 +-
 .../backup/example/HFileArchiveManager.java |   3 +-
 .../backup/example/ZKTableArchiveClient.java|   4 +-
 .../ZKSplitLogManagerCoordination.java  |  10 +-
 .../ZkSplitLogWorkerCoordination.java   |   9 +-
 .../hbase/master/ActiveMasterManager.java   |   3 +-
 .../org/apache/hadoop/hbase/master/HMaster.java |   9 +-
 .../cleaner/ReplicationZKNodeCleaner.java   |   7 +-
 .../hbase/procedure/ZKProcedureCoordinator.java |   5 +-
 .../hbase/procedure/ZKProcedureMemberRpcs.java  |  10 +-
 .../hadoop/hbase/procedure/ZKProcedureUtil.java |  19 +-
 .../hbase/regionserver/HRegionServer.java   |   3 +-
 .../security/access/ZKPermissionWatcher.java|  15 +-
 .../token/AuthenticationTokenSecretManager.java |   4 +-
 .../hbase/security/token/ZKSecretWatcher.java   |   7 +-
 .../visibility/ZKVisibilityLabelWatcher.java|   5 +-
 .../org/apache/hadoop/hbase/util/HBaseFsck.java |   3 +-
 .../hadoop/hbase/util/ZKDataMigrator.java   |   3 +-
 .../hbase/zookeeper/RegionServerTracker.java|   2 +-
 .../hbase/zookeeper/SplitOrMergeTracker.java|   4 +-
 .../hadoop/hbase/zookeeper/ZKSplitLog.java  |   4 +-
 .../apache/hadoop/hbase/GenericTestUtils.java   |   1 +
 .../hbase/client/TestMetaWithReplicas.java  |   9 +-
 .../hbase/master/TestAssignmentListener.java|   5 +-
 .../hbase/master/TestMetaShutdownHandler.java   |   3 +-
 .../hbase/master/TestTableStateManager.java |   3 +-
 .../TestCompactionInDeadRegionServer.java   |   4 +-
 .../hbase/regionserver/TestSplitLogWorker.java  |   7 +-
 .../replication/TestMasterReplication.java  |   7 +-
 .../TestReplicationStateHBaseImpl.java  |   3 +-
 .../replication/TestReplicationStateZKImpl.java |   9 +-
 .../TestReplicationTrackerZKImpl.java   |  18 +-
 .../TestReplicationSourceManager.java   |   8 +-
 .../token/TestZKSecretWatcherRefreshKeys.java   |   3 +-
 .../hadoop/hbase/zookeeper/TestZKMulti.java |  40 +--
 .../zookeeper/TestZooKeeperNodeTracker.java |   2 +-
 66 files changed, 687 insertions(+), 1044 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java

[2/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths

2017-11-10 Thread appy
http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
index 2a0764d..5954e94 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java
@@ -22,22 +22,24 @@ import static 
org.apache.hadoop.hbase.HConstants.DEFAULT_ZOOKEEPER_ZNODE_PARENT;
 import static org.apache.hadoop.hbase.HConstants.META_REPLICAS_NUM;
 import static org.apache.hadoop.hbase.HConstants.SPLIT_LOGDIR_NAME;
 import static org.apache.hadoop.hbase.HConstants.ZOOKEEPER_ZNODE_PARENT;
-import static org.apache.hadoop.hbase.HRegionInfo.DEFAULT_REPLICA_ID;
-
-import org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap;
+import static org.apache.hadoop.hbase.client.RegionInfo.DEFAULT_REPLICA_ID;
 
 import java.util.Optional;
 import java.util.stream.IntStream;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hbase.HRegionInfo;
+import org.apache.hadoop.hbase.client.RegionInfo;
 import org.apache.yetus.audience.InterfaceAudience;
 
+import org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap;
+
 /**
  * Class that hold all the paths of znode for HBase.
  */
 @InterfaceAudience.Private
 public class ZNodePaths {
+  // TODO: Replace this with ZooKeeper constant when ZOOKEEPER-277 is resolved.
+  public static final char ZNODE_PATH_SEPARATOR = '/';
 
   public final static String META_ZNODE_PREFIX = "meta-region-server";
 
@@ -90,43 +92,35 @@ public class ZNodePaths {
 baseZNode = conf.get(ZOOKEEPER_ZNODE_PARENT, 
DEFAULT_ZOOKEEPER_ZNODE_PARENT);
 ImmutableMap.Builder builder = ImmutableMap.builder();
 metaZNodePrefix = conf.get("zookeeper.znode.metaserver", 
META_ZNODE_PREFIX);
-String defaultMetaReplicaZNode = ZKUtil.joinZNode(baseZNode, 
metaZNodePrefix);
+String defaultMetaReplicaZNode = ZNodePaths.joinZNode(baseZNode, 
metaZNodePrefix);
 builder.put(DEFAULT_REPLICA_ID, defaultMetaReplicaZNode);
 int numMetaReplicas = conf.getInt(META_REPLICAS_NUM, 
DEFAULT_META_REPLICA_NUM);
 IntStream.range(1, numMetaReplicas)
 .forEachOrdered(i -> builder.put(i, defaultMetaReplicaZNode + "-" + 
i));
 metaReplicaZNodes = builder.build();
-rsZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.rs", 
"rs"));
-drainingZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.draining.rs", "draining"));
-masterAddressZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.master", "master"));
-backupMasterAddressesZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.backup.masters", "backup-masters"));
-clusterStateZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.state", "running"));
-tableZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.tableEnableDisable", "table"));
-clusterIdZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.clusterId", "hbaseid"));
-splitLogZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.splitlog", SPLIT_LOGDIR_NAME));
-balancerZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.balancer", "balancer"));
-regionNormalizerZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.regionNormalizer", "normalizer"));
-switchZNode = ZKUtil.joinZNode(baseZNode, 
conf.get("zookeeper.znode.switch", "switch"));
-tableLockZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.tableLock", "table-lock"));
-namespaceZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.namespace", "namespace"));
-masterMaintZNode = ZKUtil.joinZNode(baseZNode,
-  conf.get("zookeeper.znode.masterMaintenance", "master-maintenance"));
-replicationZNode =
-ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.replication", 
"replication"));
+rsZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.rs", "rs"));
+drainingZNode = joinZNode(baseZNode, 
conf.get("zookeeper.znode.draining.rs", "draining"));
+masterAddressZNode = joinZNode(baseZNode, 
conf.get("zookeeper.znode.master", "master"));
+backupMasterAddressesZNode =
+joinZNode(baseZNode, conf.get("zookeeper.znode.backup.masters", 
"backup-masters"));
+clusterStateZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.state", 
"running"));
+tableZNode = joinZNode(baseZNode, 
conf.get("zookeeper.znode.tableEnableDisable", "table"));
+clusterIdZNode = joinZNode(baseZNode, 
conf.get("zookeeper.znode.clusterId", "hbaseid"));
+splitLogZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.splitlog", 

hbase git commit: HBASE-16459 Remove unused hbase shell --format option

2017-11-10 Thread appy
Repository: hbase
Updated Branches:
  refs/heads/branch-1.1 f1e97d38b -> c4e3ad078


HBASE-16459 Remove unused hbase shell --format option


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c4e3ad07
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c4e3ad07
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c4e3ad07

Branch: refs/heads/branch-1.1
Commit: c4e3ad07898b42bf97d560ca1d2213276a4b9f9e
Parents: f1e97d3
Author: Apekshit Sharma 
Authored: Thu Nov 9 15:03:13 2017 -0800
Committer: Apekshit Sharma 
Committed: Fri Nov 10 09:55:11 2017 -0800

--
 bin/hirb.rb | 17 +
 1 file changed, 1 insertion(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/c4e3ad07/bin/hirb.rb
--
diff --git a/bin/hirb.rb b/bin/hirb.rb
index 0503c29..03e9af3 100644
--- a/bin/hirb.rb
+++ b/bin/hirb.rb
@@ -48,30 +48,15 @@ $LOAD_PATH.unshift Pathname.new(sources)
 cmdline_help = <

hbase git commit: HBASE-16459 Remove unused hbase shell --format option

2017-11-10 Thread appy
Repository: hbase
Updated Branches:
  refs/heads/branch-1.2 ce5264a38 -> c2d154ca6


HBASE-16459 Remove unused hbase shell --format option


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c2d154ca
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c2d154ca
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c2d154ca

Branch: refs/heads/branch-1.2
Commit: c2d154ca652289608d91098da2a8fd541c8480c2
Parents: ce5264a
Author: Apekshit Sharma 
Authored: Thu Nov 9 15:03:13 2017 -0800
Committer: Apekshit Sharma 
Committed: Fri Nov 10 09:54:54 2017 -0800

--
 bin/hirb.rb | 17 +
 1 file changed, 1 insertion(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/c2d154ca/bin/hirb.rb
--
diff --git a/bin/hirb.rb b/bin/hirb.rb
index 94b5cdb..470b093 100644
--- a/bin/hirb.rb
+++ b/bin/hirb.rb
@@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources)
 cmdline_help = <

hbase git commit: HBASE-16459 Remove unused hbase shell --format option

2017-11-10 Thread appy
Repository: hbase
Updated Branches:
  refs/heads/branch-1.3 96cad7314 -> d4acc9b68


HBASE-16459 Remove unused hbase shell --format option


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/d4acc9b6
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/d4acc9b6
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/d4acc9b6

Branch: refs/heads/branch-1.3
Commit: d4acc9b686c901f4b67a9b14966559ac12ab9f79
Parents: 96cad73
Author: Apekshit Sharma 
Authored: Thu Nov 9 15:03:13 2017 -0800
Committer: Apekshit Sharma 
Committed: Fri Nov 10 09:54:40 2017 -0800

--
 bin/hirb.rb | 17 +
 1 file changed, 1 insertion(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/d4acc9b6/bin/hirb.rb
--
diff --git a/bin/hirb.rb b/bin/hirb.rb
index 94b5cdb..470b093 100644
--- a/bin/hirb.rb
+++ b/bin/hirb.rb
@@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources)
 cmdline_help = <

hbase git commit: HBASE-16459 Remove unused hbase shell --format option

2017-11-10 Thread appy
Repository: hbase
Updated Branches:
  refs/heads/branch-1.4 6650a3ef8 -> 999cd0c3e


HBASE-16459 Remove unused hbase shell --format option


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/999cd0c3
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/999cd0c3
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/999cd0c3

Branch: refs/heads/branch-1.4
Commit: 999cd0c3e8633312115fd568e2cdef0ff83e3f96
Parents: 6650a3e
Author: Apekshit Sharma 
Authored: Thu Nov 9 15:03:13 2017 -0800
Committer: Apekshit Sharma 
Committed: Fri Nov 10 09:54:00 2017 -0800

--
 bin/hirb.rb | 17 +
 1 file changed, 1 insertion(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/999cd0c3/bin/hirb.rb
--
diff --git a/bin/hirb.rb b/bin/hirb.rb
index 2726e48..52343ca 100644
--- a/bin/hirb.rb
+++ b/bin/hirb.rb
@@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources)
 cmdline_help = <

hbase git commit: HBASE-16459 Remove unused hbase shell --format option

2017-11-10 Thread appy
Repository: hbase
Updated Branches:
  refs/heads/branch-1 696152657 -> 090ded568


HBASE-16459 Remove unused hbase shell --format option


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/090ded56
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/090ded56
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/090ded56

Branch: refs/heads/branch-1
Commit: 090ded5686a07b72f5d986c62d3b38e9c331df73
Parents: 6961526
Author: Apekshit Sharma 
Authored: Thu Nov 9 15:03:13 2017 -0800
Committer: Apekshit Sharma 
Committed: Fri Nov 10 09:52:26 2017 -0800

--
 bin/hirb.rb | 17 +
 1 file changed, 1 insertion(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/090ded56/bin/hirb.rb
--
diff --git a/bin/hirb.rb b/bin/hirb.rb
index 2726e48..52343ca 100644
--- a/bin/hirb.rb
+++ b/bin/hirb.rb
@@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources)
 cmdline_help = <

hbase git commit: HBASE-16459 Remove unused hbase shell --format option

2017-11-10 Thread appy
Repository: hbase
Updated Branches:
  refs/heads/branch-2 d757e651c -> dd19fa25d


HBASE-16459 Remove unused hbase shell --format option


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/dd19fa25
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/dd19fa25
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/dd19fa25

Branch: refs/heads/branch-2
Commit: dd19fa25d506892f1f8f620c012b2a6ae56c1e03
Parents: d757e65
Author: Apekshit Sharma 
Authored: Thu Nov 9 15:03:13 2017 -0800
Committer: Apekshit Sharma 
Committed: Fri Nov 10 00:34:16 2017 -0800

--
 bin/hirb.rb | 17 +
 1 file changed, 1 insertion(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/dd19fa25/bin/hirb.rb
--
diff --git a/bin/hirb.rb b/bin/hirb.rb
index c06c83b..35895f8 100644
--- a/bin/hirb.rb
+++ b/bin/hirb.rb
@@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources)
 cmdline_help = <

[48/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/index-all.html
--
diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html
index 0a25ebf..317fb49 100644
--- a/devapidocs/index-all.html
+++ b/devapidocs/index-all.html
@@ -4084,6 +4084,8 @@
 
 atLeast(TableSplit.Version)
 - Method in enum org.apache.hadoop.hbase.mapreduce.TableSplit.Version
 
+atomic
 - Variable in class org.apache.hadoop.hbase.regionserver.HRegion.BatchOperation
+
 ATOMIC_BOOLEAN
 - Static variable in class org.apache.hadoop.hbase.util.ClassSize
 
 Overhead for AtomicBoolean
@@ -5535,6 +5537,8 @@
 
 batchMutate(Mutation[],
 long, long) - Method in class 
org.apache.hadoop.hbase.regionserver.HRegion
 
+batchMutate(Mutation[],
 boolean, long, long) - Method in class 
org.apache.hadoop.hbase.regionserver.HRegion
+
 batchMutate(Mutation[])
 - Method in class org.apache.hadoop.hbase.regionserver.HRegion
 
 batchMutate(HRegion.BatchOperation?)
 - Method in class org.apache.hadoop.hbase.regionserver.HRegion
@@ -9104,6 +9108,8 @@
 
 cellBuilder
 - Variable in class org.apache.hadoop.hbase.codec.MessageCodec.MessageDecoder
 
+cellBuilder
 - Variable in class org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver
+
 cellBuilder
 - Variable in class org.apache.hadoop.hbase.replication.BulkLoadCellFilter
 
 CellBuilder.DataType - Enum in org.apache.hadoop.hbase
@@ -14103,6 +14109,8 @@
 
 COMPARATOR
 - Static variable in interface org.apache.hadoop.hbase.client.TableDescriptor
 
+comparator
 - Variable in class org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver
+
 comparator
 - Variable in class org.apache.hadoop.hbase.filter.CompareFilter
 
 comparator
 - Variable in class org.apache.hadoop.hbase.filter.SingleColumnValueFilter
@@ -24835,7 +24843,7 @@
 
 doBatchMutate(Mutation)
 - Method in class org.apache.hadoop.hbase.regionserver.HRegion
 
-doBatchOp(ClientProtos.RegionActionResult.Builder,
 HRegion, OperationQuota, ListClientProtos.Action, CellScanner, 
ActivePolicyEnforcement) - Method in class 
org.apache.hadoop.hbase.regionserver.RSRpcServices
+doBatchOp(ClientProtos.RegionActionResult.Builder,
 HRegion, OperationQuota, ListClientProtos.Action, CellScanner, 
ActivePolicyEnforcement, boolean) - Method in class 
org.apache.hadoop.hbase.regionserver.RSRpcServices
 
 Execute a list of Put/Delete mutations.
 
@@ -28926,6 +28934,10 @@
 
 Size of the family length field in bytes
 
+FAMILY_TO_ADD
 - Variable in class org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver
+
+FAMILY_TO_ADD_KEY
 - Static variable in class org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver
+
 familyCellMaps
 - Variable in class org.apache.hadoop.hbase.regionserver.HRegion.BatchOperation
 
 familyClose(TRegion,
 TFamily) - Method in interface org.apache.hadoop.hbase.snapshot.SnapshotManifest.RegionVisitor
@@ -29476,6 +29488,8 @@
 
 FILES
 - Static variable in class org.apache.hadoop.hbase.snapshot.SnapshotInfo.Options
 
+filesCompactedCounter
 - Variable in class org.apache.hadoop.hbase.coprocessor.example.ExampleRegionObserverWithMetrics
+
 filesCompacting
 - Variable in class org.apache.hadoop.hbase.regionserver.HStore
 
 filesGroup
 - Variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot.ExportMapper
@@ -30998,6 +31012,8 @@
 
 flushConfigTable(MapString,
 RSGroupInfo) - Method in class 
org.apache.hadoop.hbase.rsgroup.RSGroupInfoManagerImpl
 
+flushCounter
 - Variable in class org.apache.hadoop.hbase.coprocessor.example.ExampleRegionObserverWithMetrics
+
 flushed
 - Variable in class org.apache.hadoop.hbase.codec.BaseEncoder
 
 flushed
 - Variable in class org.apache.hadoop.hbase.regionserver.StoreScanner
@@ -44484,6 +44500,10 @@
 
 getRegionObserver()
 - Method in class org.apache.hadoop.hbase.coprocessor.example.ExampleRegionObserverWithMetrics
 
+getRegionObserver()
 - Method in class org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver
+
+getRegionObserver()
 - Method in class org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver
+
 getRegionObserver()
 - Method in class org.apache.hadoop.hbase.coprocessor.example.WriteHeavyIncrementObserver
 
 getRegionObserver()
 - Method in class org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver
@@ -58276,6 +58296,8 @@
 
 isAsyncPrefetch()
 - Method in class org.apache.hadoop.hbase.client.Scan
 
+isAtomic()
 - Method in class org.apache.hadoop.hbase.regionserver.HRegion.BatchOperation
+
 isAuthorizationEnabled(Connection)
 - Static method in class org.apache.hadoop.hbase.security.access.AccessControlClient
 
 Return true if authorization is supported and enabled
@@ -71914,7 +71936,7 @@
 
 MUTATION_OVERHEAD
 - Static variable in class org.apache.hadoop.hbase.client.Mutation
 
-MutationBatchOperation(HRegion,
 Mutation[], long, long) - Constructor for class 

[47/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html
index 32f538f..d9fb34d 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html
@@ -114,60 +114,64 @@
 
 
 
-org.apache.hadoop.hbase.io
+org.apache.hadoop.hbase.coprocessor.example
 
 
 
+org.apache.hadoop.hbase.io
+
+
+
 org.apache.hadoop.hbase.mapred
 
 Provides HBase http://wiki.apache.org/hadoop/HadoopMapReduce;>MapReduce
 Input/OutputFormats, a table indexing MapReduce job, and utility methods.
 
 
-
+
 org.apache.hadoop.hbase.mapreduce
 
 Provides HBase http://wiki.apache.org/hadoop/HadoopMapReduce;>MapReduce
 Input/OutputFormats, a table indexing MapReduce job, and utility methods.
 
 
-
+
 org.apache.hadoop.hbase.mapreduce.replication
 
 
-
+
 org.apache.hadoop.hbase.mob
 
 
-
+
 org.apache.hadoop.hbase.quotas
 
 
-
+
 org.apache.hadoop.hbase.regionserver
 
 
-
+
 org.apache.hadoop.hbase.regionserver.querymatcher
 
 
-
+
 org.apache.hadoop.hbase.rest.client
 
 
-
+
 org.apache.hadoop.hbase.rest.model
 
 
-
+
 org.apache.hadoop.hbase.security.access
 
 
-
+
 org.apache.hadoop.hbase.security.visibility
 
 
-
+
 org.apache.hadoop.hbase.thrift2
 
 Provides an HBase http://thrift.apache.org/;>Thrift
@@ -1202,6 +1206,25 @@ service.
 
 
 
+
+
+
+Uses of Scan in org.apache.hadoop.hbase.coprocessor.example
+
+Methods in org.apache.hadoop.hbase.coprocessor.example
 with parameters of type Scan
+
+Modifier and Type
+Method and Description
+
+
+
+void
+ScanModifyingObserver.preScannerOpen(ObserverContextRegionCoprocessorEnvironmentc,
+  Scanscan)
+
+
+
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
index 78ef84a..6955197 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
@@ -550,25 +550,25 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.client.RequestController.ReturnCode
+org.apache.hadoop.hbase.client.HBaseAdmin.ReplicationState
 org.apache.hadoop.hbase.client.IsolationLevel
+org.apache.hadoop.hbase.client.RequestController.ReturnCode
 org.apache.hadoop.hbase.client.MobCompactPartitionPolicy
+org.apache.hadoop.hbase.client.CompactionState
+org.apache.hadoop.hbase.client.Durability
+org.apache.hadoop.hbase.client.ScannerCallable.MoreResults
+org.apache.hadoop.hbase.client.AbstractResponse.ResponseType
+org.apache.hadoop.hbase.client.AsyncRequestFutureImpl.Retry
 org.apache.hadoop.hbase.client.RegionLocateType
-org.apache.hadoop.hbase.client.Consistency
-org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState
+org.apache.hadoop.hbase.client.CompactType
 org.apache.hadoop.hbase.client.AsyncProcessTask.SubmittedRows
+org.apache.hadoop.hbase.client.Consistency
 org.apache.hadoop.hbase.client.MasterSwitchType
-org.apache.hadoop.hbase.client.AbstractResponse.ResponseType
-org.apache.hadoop.hbase.client.ScannerCallable.MoreResults
-org.apache.hadoop.hbase.client.SnapshotType
+org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState
+org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState
 org.apache.hadoop.hbase.client.TableState.State
-org.apache.hadoop.hbase.client.CompactType
+org.apache.hadoop.hbase.client.SnapshotType
 org.apache.hadoop.hbase.client.Scan.ReadType
-org.apache.hadoop.hbase.client.AsyncRequestFutureImpl.Retry
-org.apache.hadoop.hbase.client.Durability
-org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState
-org.apache.hadoop.hbase.client.CompactionState
-org.apache.hadoop.hbase.client.HBaseAdmin.ReplicationState
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/client/package-use.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-use.html 
b/devapidocs/org/apache/hadoop/hbase/client/package-use.html
index 

[19/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html
index 8e0a6a2..08916af 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html
@@ -75,7 +75,7 @@
 067  };
 068
 069  protected static final State[] 
STATES_EXPECTED_ON_CLOSE = new State[] {
-070State.SPLITTING, State.SPLIT, // 
ServerCrashProcedure
+070State.SPLITTING, State.SPLIT, 
State.MERGING, // ServerCrashProcedure
 071State.OPEN,   // 
enabled/open
 072State.CLOSING // 
already in-progress (retrying)
 073  };
@@ -550,446 +550,454 @@
 542return 
!getTableRegionStates(tableName).isEmpty();
 543  }
 544
-545  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
-546return getRegionsOfTable(table, 
false);
-547  }
-548
-549  ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
-550final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
-551final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());
-552for (RegionStateNode node: nodes) {
-553  if (include(node, offline)) 
hris.add(node.getRegionInfo());
-554}
-555return hris;
-556  }
-557
-558  /**
-559   * Utility. Whether to include region 
in list of regions. Default is to
-560   * weed out split and offline 
regions.
-561   * @return True if we should include 
the codenode/code (do not include
-562   * if split or offline unless 
codeoffline/code is set to true.
-563   */
-564  boolean include(final RegionStateNode 
node, final boolean offline) {
-565if (LOG.isTraceEnabled()) {
-566  LOG.trace("WORKING ON " + node + " 
" + node.getRegionInfo());
-567}
-568if (node.isInState(State.SPLIT)) 
return false;
-569if (node.isInState(State.OFFLINE) 
 !offline) return false;
-570final RegionInfo hri = 
node.getRegionInfo();
-571return (!hri.isOffline()  
!hri.isSplit()) ||
-572((hri.isOffline() || 
hri.isSplit())  offline);
-573  }
-574
-575  /**
-576   * Returns the set of regions hosted by 
the specified server
-577   * @param serverName the server we are 
interested in
-578   * @return set of RegionInfo hosted by 
the specified server
-579   */
-580  public ListRegionInfo 
getServerRegionInfoSet(final ServerName serverName) {
-581final ServerStateNode serverInfo = 
getServerNode(serverName);
-582if (serverInfo == null) return 
Collections.emptyList();
-583
-584synchronized (serverInfo) {
-585  return 
serverInfo.getRegionInfoList();
-586}
-587  }
-588
-589  // 

-590  //  TODO: split helpers
-591  // 

-592  public void logSplit(final ServerName 
serverName) {
-593final ServerStateNode serverNode = 
getOrCreateServer(serverName);
-594synchronized (serverNode) {
-595  
serverNode.setState(ServerState.SPLITTING);
-596  /* THIS HAS TO BE WRONG. THIS IS 
SPLITTING OF REGION, NOT SPLITTING WALs.
-597  for (RegionStateNode regionNode: 
serverNode.getRegions()) {
-598synchronized (regionNode) {
-599  // TODO: Abort procedure if 
present
-600  
regionNode.setState(State.SPLITTING);
-601}
-602  }*/
-603}
-604  }
-605
-606  public void logSplit(final RegionInfo 
regionInfo) {
-607final RegionStateNode regionNode = 
getRegionNode(regionInfo);
-608synchronized (regionNode) {
-609  regionNode.setState(State.SPLIT);
-610}
-611  }
-612
-613  @VisibleForTesting
-614  public void updateRegionState(final 
RegionInfo regionInfo, final State state) {
-615final RegionStateNode regionNode = 
getOrCreateRegionNode(regionInfo);
+545  /**
+546   * @return Return online regions of 
table; does not include OFFLINE or SPLITTING regions.
+547   */
+548  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
+549return getRegionsOfTable(table, 
false);
+550  }
+551
+552  /**
+553   * @return Return the regions of the 
table; does not include OFFLINE unless you set
+554   * codeoffline/code to 
true. Does not include regions that are in the
+555   * {@link State#SPLIT} state.
+556   */
+557  public ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
+558final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
+559final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());

[45/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html
new file mode 100644
index 000..429b431
--- /dev/null
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html
@@ -0,0 +1,477 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+ScanModifyingObserver (Apache HBase 3.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods = {"i0":10,"i1":10,"i2":10};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.coprocessor.example
+Class 
ScanModifyingObserver
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver
+
+
+
+
+
+
+
+All Implemented Interfaces:
+Coprocessor, RegionCoprocessor, RegionObserver
+
+
+
+public class ScanModifyingObserver
+extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
+implements RegionCoprocessor, RegionObserver
+A RegionObserver which modifies incoming Scan requests to 
include additional
+ columns than what the user actually requested.
+
+
+
+
+
+
+
+
+
+
+
+Nested Class Summary
+
+
+
+
+Nested classes/interfaces inherited from 
interfaceorg.apache.hadoop.hbase.Coprocessor
+Coprocessor.State
+
+
+
+
+
+Nested classes/interfaces inherited from 
interfaceorg.apache.hadoop.hbase.coprocessor.RegionObserver
+RegionObserver.MutationType
+
+
+
+
+
+
+
+
+Field Summary
+
+Fields
+
+Modifier and Type
+Field and Description
+
+
+private byte[]
+FAMILY_TO_ADD
+
+
+static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+FAMILY_TO_ADD_KEY
+
+
+private byte[]
+QUALIFIER_TO_ADD
+
+
+static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+QUALIFIER_TO_ADD_KEY
+
+
+
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.Coprocessor
+PRIORITY_HIGHEST,
 PRIORITY_LOWEST,
 PRIORITY_SYSTEM,
 PRIORITY_USER,
 VERSION
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+ScanModifyingObserver()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsInstance MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalRegionObserver
+getRegionObserver()
+
+
+void
+preScannerOpen(ObserverContextRegionCoprocessorEnvironmentc,
+  Scanscan)
+Called before the client opens a new scanner.
+
+
+
+void
+start(CoprocessorEnvironmentenv)
+Called by the CoprocessorEnvironment during it's 
own startup to initialize the
+ coprocessor.
+
+
+
+
+
+
+
+Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
 title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
 title="class or interface in java.lang">hashCode, 

[36/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
index b77dc11..d1b724b 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":9,"i42":9,"i43":9,"i44":9,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":9,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":9,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":9,"i105":10,"i106":10,"i107":10,"i108":10,"i109":10
 
,"i110":10,"i111":41,"i112":41,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":9,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":42,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":9,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":10,"i175":10,"i176":9,"i177":10,"i178":10,"i179":9,"i180":9,"i181":9,"i182":9,"i183":9,"i184":9,"i185":9,"i186":9,"i187":9,"i188":10,"i189":10,"i190":10,"i191":10,"i192":10,"i193":10,"i194":10,"i195":10,"i196":10,"i197":9,"i198":10,"i199":10,"i200":10,"i201":10,"i202":10,"i203":10,"i204":10,"i205":10,"i206":10,"i207":10,"i208":10,"i209":10,"i210":10
 
,"i211":10,"i212":10,"i213":10,"i214":10,"i215":10,"i216":10,"i217":10,"i218":10,"i219":10,"i220":10,"i221":10,"i222":10,"i223":10,"i224":10,"i225":10,"i226":10,"i227":10,"i228":9,"i229":9,"i230":10,"i231":10,"i232":10,"i233":10,"i234":10,"i235":10,"i236":10,"i237":10,"i238":10,"i239":10,"i240":10,"i241":9,"i242":10,"i243":10,"i244":10,"i245":10,"i246":10,"i247":10,"i248":10,"i249":10,"i250":10,"i251":10,"i252":10,"i253":10,"i254":10,"i255":10,"i256":9,"i257":10,"i258":10,"i259":10,"i260":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":9,"i43":9,"i44":9,"i45":9,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":9,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":9,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":9,"i106":10,"i107":10,"i108":10,"i109":10
 

[46/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
index 1d1a3ec..9fd977c 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = {"i0":10,"i1":10,"i2":10,"i3":10};
+var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-class ExampleRegionObserverWithMetrics.ExampleRegionObserver
+class ExampleRegionObserverWithMetrics.ExampleRegionObserver
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements RegionCoprocessor, RegionObserver
 
@@ -202,13 +202,40 @@ implements 
 void
+postCompactSelection(ObserverContextRegionCoprocessorEnvironmentc,
+Storestore,
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends StoreFileselected,
+CompactionLifeCycleTrackertracker,
+CompactionRequestrequest)
+Called after the StoreFiles to compact 
have been selected from the available
+ candidates.
+
+
+
+void
+postFlush(ObserverContextRegionCoprocessorEnvironmentc,
+ FlushLifeCycleTrackertracker)
+Called after the memstore is flushed to disk.
+
+
+
+void
+postFlush(ObserverContextRegionCoprocessorEnvironmentc,
+ Storestore,
+ StoreFileresultFile,
+ FlushLifeCycleTrackertracker)
+Called after a Store's memstore is flushed to disk.
+
+
+
+void
 postGetOp(ObserverContextRegionCoprocessorEnvironmente,
  Getget,
  http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListCellresults)
 Called after the client performs a Get
 
 
-
+
 void
 preGetOp(ObserverContextRegionCoprocessorEnvironmente,
 Getget,
@@ -243,7 +270,7 @@ implements RegionObserver
-postAppend,
 postBatchMutate,
 postBatchMutateIndispensably,
 postBulkLoadHFile,
 postCheckAndDelete,
 postCheckAndPut,
 postClose,
 postCloseRegionOperation, postCommitStoreFile,
 postCompact,
 postCompactSelection,
 postDelete,
 postExists,
 postFlush,
 postFlush,
 postIncrement,
 postInstantiateDeleteTracker,
 postMemStoreCompaction,
 postMutationBeforeWAL,
 postOpen,
 postPut,
 postReplayWALs,
 postScannerClose,
 postScannerFilterRow,
 postScannerNext,
 postScannerOpen,
 postStartRegionOperation,
 postStoreFileReaderOpen,
 postWALRestore,
 preAppend, 
preAppendAfterRowLock,
 preBatchMutate,
 preBulkLoadHFile,
 preCheckAndDelete, preCheckAndDeleteAfterRowLock,
 preCheckAndPut,
 preCheckAndPutAfterRowLock,
 preClose,
 preCommitStoreFile,
 preCompact,
 preCompactScannerOpen,
 preCompactSelection,
 preDelete,
 preExists,
 preFlush,
 preFlush,
 preFlushScannerOpen, preIncrement,
 preIncrementAfterRowLock,
 preMemStoreCompaction,
 preMemStoreCompactionCompact, preMemStoreCompactionCompactScannerOpen,
 preOpen,
 prePrepareTimeStampForDeleteVersion,
 prePut,
 preReplayWALs,
 preScannerClose,
 preScannerNext,
 preScannerOpen, preStoreFileReaderOpen,
 preStoreScannerOpen,
 preWALRestore
+postAppend,
 postBatchMutate,
 postBatchMutateIndispensably,
 postBulkLoadHFile,
 postCheckAndDelete,
 postCheckAndPut,
 postClose,
 postCloseRegionOperation, postCommitStoreFile,
 postCompact,
 postDelete,
 postExists,
 postIncrement,
 postInstantiateDeleteTracker,
 postMemStoreCompaction,
 postMutationBeforeWAL,
 postOpen,
 postPut,
 postReplayWALs,
 postScannerClose,
 postScannerFilterRow,
 postScannerNext,
 postScannerOpen,
 postStartRegionOperation,
 postStoreFileReaderOpen,
 postWALRestore,
 preAppend,
 preAppendAfterRowLock,
 preBatchMutate,
 preBulkLoadHFile,
 preCheckAndDelete,
 preCheckAndDeleteAfterRowLock,
 preCheckAndPut,
 preCheckAndPutAfterRowLock,
 preClose,
 preCommitStoreFile,
 pre
 Compact, preCompactScannerOpen,
 preCompactSelection,
 preDelete, 
preExists,
 preFlush,
 preFlush,
 preFlushScannerOpen,
 preIncrement,
 preIncrementAfterRowLock,
 preMemStoreCompaction,
 preMemStoreCompactionCompact,
 preMemStoreCompactionCompactScannerOpen,
 preOpen,
 prePrepareTimeStampForDeleteVersion,
 prePut,
 

[44/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html
new file mode 100644
index 000..075273c
--- /dev/null
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html
@@ -0,0 +1,515 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+ValueRewritingObserver (Apache HBase 3.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods = {"i0":10,"i1":10,"i2":10};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.coprocessor.example
+Class 
ValueRewritingObserver
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver
+
+
+
+
+
+
+
+All Implemented Interfaces:
+Coprocessor, RegionCoprocessor, RegionObserver
+
+
+
+public class ValueRewritingObserver
+extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
+implements RegionObserver, 
RegionCoprocessor
+This RegionObserver replaces the values of Puts from one 
value to another on compaction.
+
+
+
+
+
+
+
+
+
+
+
+Nested Class Summary
+
+
+
+
+Nested classes/interfaces inherited from 
interfaceorg.apache.hadoop.hbase.coprocessor.RegionObserver
+RegionObserver.MutationType
+
+
+
+
+
+Nested classes/interfaces inherited from 
interfaceorg.apache.hadoop.hbase.Coprocessor
+Coprocessor.State
+
+
+
+
+
+
+
+
+Field Summary
+
+Fields
+
+Modifier and Type
+Field and Description
+
+
+private CellBuilder
+cellBuilder
+
+
+private Bytes.ByteArrayComparator
+comparator
+
+
+static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+ORIGINAL_VALUE_KEY
+
+
+static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
+REPLACED_VALUE_KEY
+
+
+private byte[]
+replacedValue
+
+
+private byte[]
+sourceValue
+
+
+
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.Coprocessor
+PRIORITY_HIGHEST,
 PRIORITY_LOWEST,
 PRIORITY_SYSTEM,
 PRIORITY_USER,
 VERSION
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+ValueRewritingObserver()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsInstance MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">OptionalRegionObserver
+getRegionObserver()
+
+
+InternalScanner
+preCompact(ObserverContextRegionCoprocessorEnvironmentc,
+  Storestore,
+  InternalScannerscanner,
+  ScanTypescanType,
+  CompactionLifeCycleTrackertracker,
+  CompactionRequestrequest)
+Called prior to writing the StoreFiles selected for 
compaction into a new
+ StoreFile.
+
+
+
+void
+start(CoprocessorEnvironmentenv)
+Called by the CoprocessorEnvironment during it's 
own startup to initialize the
+ coprocessor.
+
+
+
+
+
+
+
+Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, 

[37/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
index 5f09cd0..5833858 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public static class HRegion.RowLockImpl
+public static class HRegion.RowLockImpl
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements Region.RowLock
 Class used to represent a lock on a row.
@@ -226,7 +226,7 @@ implements 
 
 context
-private finalHRegion.RowLockContext context
+private finalHRegion.RowLockContext context
 
 
 
@@ -235,7 +235,7 @@ implements 
 
 lock
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true;
 title="class or interface in java.util.concurrent.locks">Lock lock
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true;
 title="class or interface in java.util.concurrent.locks">Lock lock
 
 
 
@@ -252,7 +252,7 @@ implements 
 
 RowLockImpl
-publicRowLockImpl(HRegion.RowLockContextcontext,
+publicRowLockImpl(HRegion.RowLockContextcontext,
http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true;
 title="class or interface in 
java.util.concurrent.locks">Locklock)
 
 
@@ -270,7 +270,7 @@ implements 
 
 getLock
-publichttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true;
 title="class or interface in java.util.concurrent.locks">LockgetLock()
+publichttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true;
 title="class or interface in java.util.concurrent.locks">LockgetLock()
 
 
 
@@ -279,7 +279,7 @@ implements 
 
 getContext
-publicHRegion.RowLockContextgetContext()
+publicHRegion.RowLockContextgetContext()
 
 
 
@@ -288,7 +288,7 @@ implements 
 
 release
-publicvoidrelease()
+publicvoidrelease()
 Description copied from 
interface:Region.RowLock
 Release the given lock.  If there are no remaining locks 
held by the current thread
  then unlock the row and allow other threads to acquire the lock.
@@ -304,7 +304,7 @@ implements 
 
 toString
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtoString()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringtoString()
 
 Overrides:
 http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
 title="class or interface in java.lang">toStringin 
classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object



[50/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/checkstyle-aggregate.html
--
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index 3d0abd1..3fa0a8f 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  Checkstyle Results
 
@@ -286,10 +286,10 @@
 Warnings
 Errors
 
-3422
+3426
 0
 0
-21453
+21437
 
 Files
 
@@ -6262,7 +6262,7 @@
 org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java
 0
 0
-84
+75
 
 org/apache/hadoop/hbase/master/assignment/MockMasterServices.java
 0
@@ -6277,7 +6277,7 @@
 org/apache/hadoop/hbase/master/assignment/RegionStateStore.java
 0
 0
-10
+9
 
 org/apache/hadoop/hbase/master/assignment/RegionStates.java
 0
@@ -8377,7 +8377,7 @@
 org/apache/hadoop/hbase/regionserver/RSRpcServices.java
 0
 0
-142
+136
 
 org/apache/hadoop/hbase/regionserver/RSStatusServlet.java
 0
@@ -13199,12 +13199,12 @@
 
 
 http://checkstyle.sourceforge.net/config_blocks.html#NeedBraces;>NeedBraces
-2281
+2280
 Error
 
 coding
 http://checkstyle.sourceforge.net/config_coding.html#EmptyStatement;>EmptyStatement
-56
+55
 Error
 
 
@@ -13284,7 +13284,7 @@
 caseIndent: 2
 basicOffset: 2
 lineWrappingIndentation: 2
-6309
+6300
 Error
 
 javadoc
@@ -13296,7 +13296,7 @@
 
 
 http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription;>NonEmptyAtclauseDescription
-4355
+4351
 Error
 
 misc
@@ -13314,7 +13314,7 @@
 
 max: 100
 ignorePattern: ^package.*|^import.*|a 
href|href|http://|https://|ftp://|org.apache.thrift.|com.google.protobuf.|hbase.protobuf.generated
-1952
+1951
 Error
 
 
@@ -84053,91 +84053,91 @@
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-1327
+1336
 
 Error
 javadoc
 NonEmptyAtclauseDescription
 At-clause should have a non-empty description.
-1345
+1354
 
 Error
 javadoc
 NonEmptyAtclauseDescription
 At-clause should have a non-empty description.
-1347
+1356
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-1350
+1359
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-1355
+1364
 
 Error
 javadoc
 NonEmptyAtclauseDescription
 At-clause should have a non-empty description.
-1363
+1372
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-1368
+1377
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-1373
+1382
 
 Error
 blocks
 NeedBraces
 'else' construct must use '{}'s.
-1374
+1383
 
 Error
 blocks
 EmptyBlock
 Must have at least one statement.
-1455
+1464
 
 Error
 sizes
 LineLength
 Line is longer than 100 characters (found 101).
-1482
+1491
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-1638
+1652
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-1765
+1779
 
 Error
 indentation
 Indentation
 'lambda arguments' have incorrect indentation level 8, expected level 
should be 6.
-1810
+1824
 
 Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-1827
+1841
 
 org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.java
 
@@ -84452,515 +84452,461 @@
 imports
 ImportOrder
 Wrong order for 
'org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting'
 import.
-67
+68
 
 Error
 indentation
 Indentation
 'case' child have incorrect indentation level 6, expected level should be 
8.
-216
+217
 
 Error
 indentation
 Indentation
 'if' have incorrect indentation level 8, expected level should be 10.
-217
+218
 
 Error
 indentation
 Indentation
 'if' child have incorrect indentation level 10, expected level should be 
12.
-219
+220
 
 Error
 indentation
 Indentation
 'if rcurly' have incorrect indentation level 8, expected level should be 
10.
-220
-
-Error
-indentation
-Indentation
-'block' child have incorrect indentation level 8, expected level should be 
10.
 221
-
-Error
-indentation
-Indentation
-'block' child have incorrect indentation level 8, expected level should be 
10.
-222
-
-Error
-indentation
-Indentation
-'case' child have incorrect indentation level 6, expected level should be 
8.
-223
-
-Error
-indentation
-Indentation
-'block' child have incorrect indentation level 8, expected level should be 
10.
-224
 
 Error
 indentation
 Indentation
 'block' child have incorrect indentation level 8, expected level should be 
10.
-225
+222
 
 Error
 indentation
 Indentation
 'block' child have incorrect indentation level 8, expected level should be 
10.
-226
+223
 
 Error
 indentation
 Indentation
 'case' child have incorrect indentation level 6, expected level should be 
8.
-227
+224
 
 Error
 indentation
 Indentation
 'block' child have incorrect indentation level 8, expected level should be 
10.
-228
+225
 
 Error
 indentation
 Indentation
 'block' child have incorrect indentation level 8, expected level should be 
10.
-229
+226
 
 Error
 indentation
 Indentation
 'block' child have incorrect indentation level 8, expected level should be 
10.
-230
+227
 
 Error
 

[42/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
index a749d86..c80374f 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
@@ -129,7 +129,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class MergeTableRegionsProcedure
+public class MergeTableRegionsProcedure
 extends AbstractStateMachineTableProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsState
 The procedure to Merge a region in a table.
  This procedure takes an exclusive table lock since it is working over 
multiple regions.
@@ -483,15 +483,16 @@ extends 
-void
-setRegionStateToMerging(MasterProcedureEnvenv)
-Set the region states to MERGING state
+private void
+setRegionStateBackToOpen(MasterProcedureEnvenv)
+Rollback the region state change
+ Not used for now, since rollbackCloseRegionsForMerge() will mark regions as 
OPEN
 
 
 
-private void
-setRegionStateToRevertMerging(MasterProcedureEnvenv)
-Rollback the region state change
+void
+setRegionStateToMerging(MasterProcedureEnvenv)
+Set the region states to MERGING state
 
 
 
@@ -556,7 +557,7 @@ extends 
 
 LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static finalorg.apache.commons.logging.Log LOG
 
 
 
@@ -565,7 +566,7 @@ extends 
 
 traceEnabled
-privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
+privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
 
 
 
@@ -574,7 +575,7 @@ extends 
 
 lock
-private volatileboolean lock
+private volatileboolean lock
 
 
 
@@ -583,7 +584,7 @@ extends 
 
 regionLocation
-privateServerName regionLocation
+privateServerName regionLocation
 
 
 
@@ -592,7 +593,7 @@ extends 
 
 regionsToMerge
-privateRegionInfo[] regionsToMerge
+privateRegionInfo[] regionsToMerge
 
 
 
@@ -601,7 +602,7 @@ extends 
 
 mergedRegion
-privateRegionInfo mergedRegion
+privateRegionInfo mergedRegion
 
 
 
@@ -610,7 +611,7 @@ extends 
 
 forcible
-privateboolean forcible
+privateboolean forcible
 
 
 
@@ -627,7 +628,7 @@ extends 
 
 MergeTableRegionsProcedure
-publicMergeTableRegionsProcedure()
+publicMergeTableRegionsProcedure()
 
 
 
@@ -636,7 +637,7 @@ extends 
 
 MergeTableRegionsProcedure
-publicMergeTableRegionsProcedure(MasterProcedureEnvenv,
+publicMergeTableRegionsProcedure(MasterProcedureEnvenv,
   RegionInforegionToMergeA,
   RegionInforegionToMergeB)
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
@@ -652,7 +653,7 @@ extends 
 
 MergeTableRegionsProcedure
-publicMergeTableRegionsProcedure(MasterProcedureEnvenv,
+publicMergeTableRegionsProcedure(MasterProcedureEnvenv,
   RegionInforegionToMergeA,
   RegionInforegionToMergeB,
   booleanforcible)
@@ -669,7 +670,7 @@ extends 
 
 MergeTableRegionsProcedure
-publicMergeTableRegionsProcedure(MasterProcedureEnvenv,
+publicMergeTableRegionsProcedure(MasterProcedureEnvenv,
   RegionInfo[]regionsToMerge,
   booleanforcible)
throws MergeRegionException
@@ -693,7 +694,7 @@ extends 
 
 checkRegionsToMerge
-private staticvoidcheckRegionsToMerge(RegionInfo[]regionsToMerge,
+private staticvoidcheckRegionsToMerge(RegionInfo[]regionsToMerge,
 booleanforcible)
  throws MergeRegionException
 
@@ -708,7 +709,7 @@ extends 
 
 checkRegionsToMerge
-private staticvoidcheckRegionsToMerge(RegionInforegionToMergeA,
+private staticvoidcheckRegionsToMerge(RegionInforegionToMergeA,
 RegionInforegionToMergeB,
 booleanforcible)
  throws MergeRegionException
@@ -724,7 +725,7 @@ extends 
 
 createMergedRegionInfo
-private staticRegionInfocreateMergedRegionInfo(RegionInfo[]regionsToMerge)
+private staticRegionInfocreateMergedRegionInfo(RegionInfo[]regionsToMerge)
 
 
 
@@ -733,7 +734,7 @@ extends 
 
 createMergedRegionInfo
-private staticRegionInfocreateMergedRegionInfo(RegionInforegionToMergeA,
+private 

hbase-site git commit: INFRA-10751 Empty commit

2017-11-10 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site a108018f1 -> bac2e663f


INFRA-10751 Empty commit


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/bac2e663
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/bac2e663
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/bac2e663

Branch: refs/heads/asf-site
Commit: bac2e663fcd1745673991e367bab24750e970ab6
Parents: a108018
Author: jenkins 
Authored: Fri Nov 10 15:30:07 2017 +
Committer: jenkins 
Committed: Fri Nov 10 15:30:07 2017 +

--

--




[43/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
index d4bc51c..efa1c09 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
@@ -406,7 +406,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in 
java.util">Setbyte[]regionNames)
 
 
-void
+(package private) void
 checkOnlineRegionsReportForMeta(RegionStates.ServerStateNodeserverNode,
http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in 
java.util">Setbyte[]regionNames)
 
@@ -1299,7 +1299,7 @@ implements 
 
 pendingAssignQueue
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true;
 title="class or interface in java.util">ArrayListRegionStates.RegionStateNode 
pendingAssignQueue
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true;
 title="class or interface in java.util">ArrayListRegionStates.RegionStateNode 
pendingAssignQueue
 
 
 
@@ -1308,7 +1308,7 @@ implements 
 
 assignQueueLock
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/ReentrantLock.html?is-external=true;
 title="class or interface in java.util.concurrent.locks">ReentrantLock assignQueueLock
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/ReentrantLock.html?is-external=true;
 title="class or interface in java.util.concurrent.locks">ReentrantLock assignQueueLock
 
 
 
@@ -1317,7 +1317,7 @@ implements 
 
 assignQueueFullCond
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Condition.html?is-external=true;
 title="class or interface in java.util.concurrent.locks">Condition assignQueueFullCond
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Condition.html?is-external=true;
 title="class or interface in java.util.concurrent.locks">Condition assignQueueFullCond
 
 
 
@@ -2126,8 +2126,8 @@ implements 
 
 checkOnlineRegionsReportForMeta
-publicvoidcheckOnlineRegionsReportForMeta(RegionStates.ServerStateNodeserverNode,
-http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in 
java.util">Setbyte[]regionNames)
+voidcheckOnlineRegionsReportForMeta(RegionStates.ServerStateNodeserverNode,
+ http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in 
java.util">Setbyte[]regionNames)
 
 
 
@@ -2225,7 +2225,7 @@ implements 
 
 processofflineServersWithOnlineRegions
-privatebooleanprocessofflineServersWithOnlineRegions()
+privatebooleanprocessofflineServersWithOnlineRegions()
 
 
 
@@ -2234,7 +2234,7 @@ implements 
 
 isFailoverCleanupDone
-publicbooleanisFailoverCleanupDone()
+publicbooleanisFailoverCleanupDone()
 Used by ServerCrashProcedure to make sure AssignmentManager 
has completed
  the failover cleanup before re-assigning regions of dead servers. So that
  when re-assignment happens, AssignmentManager has proper region states.
@@ -2246,7 +2246,7 @@ implements 
 
 setFailoverCleanupDone
-publicvoidsetFailoverCleanupDone(booleanb)
+publicvoidsetFailoverCleanupDone(booleanb)
 Used by ServerCrashProcedure tests verify the ability to 
suspend the
  execution of the ServerCrashProcedure.
 
@@ -2257,7 +2257,7 @@ implements 
 
 getFailoverCleanupEvent
-publicProcedureEventgetFailoverCleanupEvent()
+publicProcedureEventgetFailoverCleanupEvent()
 
 
 
@@ -2266,7 +2266,7 @@ implements 
 
 checkFailoverCleanupCompleted
-privatevoidcheckFailoverCleanupCompleted(RegionInfohri)
+privatevoidcheckFailoverCleanupCompleted(RegionInfohri)
 throws PleaseHoldException
 Used to check if the failover cleanup is done.
  if not we throw PleaseHoldException since we are rebuilding the 
RegionStates
@@ -2284,7 +2284,7 @@ implements 
 
 getNumRegionsOpened
-publicintgetNumRegionsOpened()
+publicintgetNumRegionsOpened()
 
 
 
@@ -2293,7 +2293,7 @@ implements 
 
 submitServerCrash
-publicvoidsubmitServerCrash(ServerNameserverName,
+publicvoidsubmitServerCrash(ServerNameserverName,
   booleanshouldSplitWal)
 
 
@@ -2303,7 +2303,7 @@ implements 
 
 offlineRegion
-publicvoidofflineRegion(RegionInforegionInfo)
+publicvoidofflineRegion(RegionInforegionInfo)
 
 
 
@@ -2312,7 +2312,7 @@ implements 
 
 onlineRegion
-publicvoidonlineRegion(RegionInforegionInfo,

[49/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index ad53a23..e3b0bd0 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
 en-us
 2007 - 2017 The Apache Software Foundation
 
-  File: 3422,
- Errors: 21453,
+  File: 3426,
+ Errors: 21437,
  Warnings: 0,
  Infos: 0
   
@@ -7704,6 +7704,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.example.TestScanModifyingObserver.java;>org/apache/hadoop/hbase/coprocessor/example/TestScanModifyingObserver.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.AbstractByteRange.java;>org/apache/hadoop/hbase/util/AbstractByteRange.java
 
 
@@ -14942,6 +14956,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver.java;>org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.CommonFSUtils.java;>org/apache/hadoop/hbase/util/CommonFSUtils.java
 
 
@@ -18339,7 +18367,7 @@ under the License.
   0
 
 
-  142
+  136
 
   
   
@@ -20789,7 +20817,7 @@ under the License.
   0
 
 
-  10
+  9
 
   
   
@@ -21349,7 +21377,7 @@ under the License.
   0
 
 
-  84
+  75
 
   
   
@@ -35018,6 +35046,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.example.TestValueReplacingCompaction.java;>org/apache/hadoop/hbase/coprocessor/example/TestValueReplacingCompaction.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.StripeStoreConfig.java;>org/apache/hadoop/hbase/regionserver/StripeStoreConfig.java
 
 
@@ -40534,6 +40576,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver.java;>org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.client.example.RefreshHFilesClient.java;>org/apache/hadoop/hbase/client/example/RefreshHFilesClient.java
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/coc.html
--
diff --git a/coc.html b/coc.html
index 51ce016..14a8353 100644
--- a/coc.html
+++ b/coc.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase  
   Code of Conduct Policy
@@ -380,7 +380,7 @@ email to mailto:priv...@hbase.apache.org;>the priv
 https://www.apache.org/;>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-11-09
+  Last Published: 
2017-11-10
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/cygwin.html

[38/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
index 24779ca..e04d11d 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
@@ -121,7 +121,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-class HRegion.RegionScannerImpl
+class HRegion.RegionScannerImpl
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements RegionScanner, Shipper, RpcCallback
 RegionScannerImpl is used to combine scanners from multiple 
Stores (aka column families).
@@ -425,7 +425,7 @@ implements 
 
 storeHeap
-KeyValueHeap storeHeap
+KeyValueHeap storeHeap
 
 
 
@@ -434,7 +434,7 @@ implements 
 
 joinedHeap
-KeyValueHeap joinedHeap
+KeyValueHeap joinedHeap
 Heap of key-values that are not essential for the provided 
filters and are thus read
  on demand, if on-demand column family loading is enabled.
 
@@ -445,7 +445,7 @@ implements 
 
 joinedContinuationRow
-protectedCell joinedContinuationRow
+protectedCell joinedContinuationRow
 If the joined heap data gathering is interrupted due to 
scan limits, this will
  contain the row for which we are populating the values.
 
@@ -456,7 +456,7 @@ implements 
 
 filterClosed
-privateboolean filterClosed
+privateboolean filterClosed
 
 
 
@@ -465,7 +465,7 @@ implements 
 
 stopRow
-protected finalbyte[] stopRow
+protected finalbyte[] stopRow
 
 
 
@@ -474,7 +474,7 @@ implements 
 
 includeStopRow
-protected finalboolean includeStopRow
+protected finalboolean includeStopRow
 
 
 
@@ -483,7 +483,7 @@ implements 
 
 region
-protected finalHRegion region
+protected finalHRegion region
 
 
 
@@ -492,7 +492,7 @@ implements 
 
 comparator
-protected finalCellComparator comparator
+protected finalCellComparator comparator
 
 
 
@@ -501,7 +501,7 @@ implements 
 
 readPt
-private finallong readPt
+private finallong readPt
 
 
 
@@ -510,7 +510,7 @@ implements 
 
 maxResultSize
-private finallong maxResultSize
+private finallong maxResultSize
 
 
 
@@ -519,7 +519,7 @@ implements 
 
 defaultScannerContext
-private finalScannerContext defaultScannerContext
+private finalScannerContext defaultScannerContext
 
 
 
@@ -528,7 +528,7 @@ implements 
 
 filter
-private finalFilterWrapper filter
+private finalFilterWrapper filter
 
 
 
@@ -545,7 +545,7 @@ implements 
 
 RegionScannerImpl
-RegionScannerImpl(Scanscan,
+RegionScannerImpl(Scanscan,
   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListKeyValueScanneradditionalScanners,
   HRegionregion)
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
@@ -561,7 +561,7 @@ implements 
 
 RegionScannerImpl
-RegionScannerImpl(Scanscan,
+RegionScannerImpl(Scanscan,
   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListKeyValueScanneradditionalScanners,
   HRegionregion,
   longnonceGroup,
@@ -587,7 +587,7 @@ implements 
 
 getRegionInfo
-publicRegionInfogetRegionInfo()
+publicRegionInfogetRegionInfo()
 
 Specified by:
 getRegionInfoin
 interfaceRegionScanner
@@ -602,7 +602,7 @@ implements 
 
 initializeScanners
-protectedvoidinitializeScanners(Scanscan,
+protectedvoidinitializeScanners(Scanscan,
   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListKeyValueScanneradditionalScanners)
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
@@ -617,7 +617,7 @@ implements 
 
 initializeKVHeap
-protectedvoidinitializeKVHeap(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListKeyValueScannerscanners,
+protectedvoidinitializeKVHeap(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListKeyValueScannerscanners,
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListKeyValueScannerjoinedScanners,
 HRegionregion)
  throws 

[32/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html
index dc2f127..91776e4 100644
--- a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html
@@ -110,8 +110,8 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.ByteArrayComparableModel.ComparatorType
 org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.FilterType
+org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.ByteArrayComparableModel.ComparatorType
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html
index d2946f4..793c33c 100644
--- a/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html
@@ -136,9 +136,9 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.security.access.AccessController.OpType
-org.apache.hadoop.hbase.security.access.Permission.Action
 org.apache.hadoop.hbase.security.access.AccessControlFilter.Strategy
+org.apache.hadoop.hbase.security.access.Permission.Action
+org.apache.hadoop.hbase.security.access.AccessController.OpType
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
index c4d3f6e..bec1bbc 100644
--- a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html
@@ -191,8 +191,8 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection
 org.apache.hadoop.hbase.security.AuthMethod
+org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection
 org.apache.hadoop.hbase.security.SaslStatus
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
index 29f4bfc..59ab0ce 100644
--- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html
@@ -198,9 +198,9 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType
 org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage
 org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType

[27/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
index cb99dd4..e7374e1 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
@@ -55,739 +55,744 @@
 047import 
org.apache.hadoop.hbase.master.MasterCoprocessorHost;
 048import 
org.apache.hadoop.hbase.master.MasterFileSystem;
 049import 
org.apache.hadoop.hbase.master.RegionState;
-050import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
-051import 
org.apache.hadoop.hbase.master.procedure.AbstractStateMachineTableProcedure;
-052import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
-053import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
-054import 
org.apache.hadoop.hbase.procedure2.ProcedureMetrics;
-055import 
org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer;
-056import 
org.apache.hadoop.hbase.procedure2.ProcedureSuspendedException;
-057import 
org.apache.hadoop.hbase.procedure2.ProcedureYieldException;
-058import 
org.apache.hadoop.hbase.quotas.QuotaExceededException;
-059import 
org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
-060import 
org.apache.hadoop.hbase.regionserver.HStoreFile;
-061import 
org.apache.hadoop.hbase.regionserver.StoreFileInfo;
-062import 
org.apache.hadoop.hbase.util.Bytes;
-063import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-064import 
org.apache.hadoop.hbase.util.FSUtils;
-065import 
org.apache.yetus.audience.InterfaceAudience;
-066
-067import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-068import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-069import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
-070import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
-071import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsState;
-072
-073/**
-074 * The procedure to Merge a region in a 
table.
-075 * This procedure takes an exclusive 
table lock since it is working over multiple regions.
-076 * It holds the lock for the life of the 
procedure.
-077 */
-078@InterfaceAudience.Private
-079public class MergeTableRegionsProcedure
-080extends 
AbstractStateMachineTableProcedureMergeTableRegionsState {
-081  private static final Log LOG = 
LogFactory.getLog(MergeTableRegionsProcedure.class);
-082  private Boolean traceEnabled;
-083  private volatile boolean lock = 
false;
-084  private ServerName regionLocation;
-085  private RegionInfo[] regionsToMerge;
-086  private RegionInfo mergedRegion;
-087  private boolean forcible;
-088
-089  public MergeTableRegionsProcedure() {
-090// Required by the Procedure 
framework to create the procedure on replay
-091  }
-092
-093  public MergeTableRegionsProcedure(final 
MasterProcedureEnv env,
-094  final RegionInfo regionToMergeA, 
final RegionInfo regionToMergeB) throws IOException {
-095this(env, regionToMergeA, 
regionToMergeB, false);
-096  }
-097
-098  public MergeTableRegionsProcedure(final 
MasterProcedureEnv env,
-099  final RegionInfo regionToMergeA, 
final RegionInfo regionToMergeB,
-100  final boolean forcible) throws 
MergeRegionException {
-101this(env, new RegionInfo[] 
{regionToMergeA, regionToMergeB}, forcible);
-102  }
-103
-104  public MergeTableRegionsProcedure(final 
MasterProcedureEnv env,
-105  final RegionInfo[] regionsToMerge, 
final boolean forcible)
-106  throws MergeRegionException {
-107super(env);
-108
-109// Check daughter regions and make 
sure that we have valid daughter regions
-110// before doing the real work.
-111checkRegionsToMerge(regionsToMerge, 
forcible);
-112
-113// WARN: make sure there is no parent 
region of the two merging regions in
-114// hbase:meta If exists, fixing up 
daughters would cause daughter regions(we
-115// have merged one) online again when 
we restart master, so we should clear
-116// the parent region to prevent the 
above case
-117// Since HBASE-7721, we don't need 
fix up daughters any more. so here do nothing
-118this.regionsToMerge = 
regionsToMerge;
-119this.mergedRegion = 
createMergedRegionInfo(regionsToMerge);
-120this.forcible = forcible;
-121  }
-122
-123  private static void 
checkRegionsToMerge(final RegionInfo[] regionsToMerge,
-124  final boolean forcible) throws 
MergeRegionException {
-125// For now, we only merge 2 
regions.
-126// It could be extended to more than 
2 

[30/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
index f0775d0..40cd159 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html
@@ -943,7 +943,7 @@
 935wakeServerReportEvent(serverNode);
 936  }
 937
-938  public void 
checkOnlineRegionsReportForMeta(final ServerStateNode serverNode,
+938  void 
checkOnlineRegionsReportForMeta(final ServerStateNode serverNode,
 939  final Setbyte[] 
regionNames) {
 940try {
 941  for (byte[] regionName: 
regionNames) {
@@ -959,7 +959,7 @@
 951final RegionStateNode regionNode 
= regionStates.getOrCreateRegionNode(hri);
 952LOG.info("META REPORTED: " + 
regionNode);
 953if (!reportTransition(regionNode, 
serverNode, TransitionCode.OPENED, 0)) {
-954  LOG.warn("META REPORTED but no 
procedure found");
+954  LOG.warn("META REPORTED but no 
procedure found (complete?)");
 955  
regionNode.setRegionLocation(serverNode.getServerName());
 956} else if (LOG.isTraceEnabled()) 
{
 957  LOG.trace("META REPORTED: " + 
regionNode);
@@ -1191,671 +1191,685 @@
 1183  public void visitRegionState(final 
RegionInfo regionInfo, final State state,
 1184  final ServerName 
regionLocation, final ServerName lastHost, final long openSeqNum) {
 1185final RegionStateNode regionNode 
= regionStates.getOrCreateRegionNode(regionInfo);
-1186synchronized (regionNode) {
-1187  if 
(!regionNode.isInTransition()) {
-1188
regionNode.setState(state);
-1189
regionNode.setLastHost(lastHost);
-1190
regionNode.setRegionLocation(regionLocation);
-1191
regionNode.setOpenSeqNum(openSeqNum);
-1192
-1193if (state == State.OPEN) {
-1194  assert regionLocation != 
null : "found null region location for " + regionNode;
-1195  
regionStates.addRegionToServer(regionLocation, regionNode);
-1196} else if (state == 
State.OFFLINE || regionInfo.isOffline()) {
-1197  
regionStates.addToOfflineRegions(regionNode);
-1198} else {
-1199  // These regions should 
have a procedure in replay
-1200  
regionStates.addRegionInTransition(regionNode, null);
-1201}
-1202  }
-1203}
-1204  }
-1205});
-1206
-1207// every assignment is blocked until 
meta is loaded.
-1208wakeMetaLoadedEvent();
-1209  }
-1210
-1211  // TODO: the assumption here is that 
if RSs are crashing while we are executing this
-1212  // they will be handled by the SSH 
that are put in the ServerManager "queue".
-1213  // we can integrate this a bit 
better.
-1214  private boolean 
processofflineServersWithOnlineRegions() {
-1215boolean failover = 
!master.getServerManager().getDeadServers().isEmpty();
-1216
-1217final SetServerName 
offlineServersWithOnlineRegions = new HashSetServerName();
-1218final ArrayListRegionInfo 
regionsToAssign = new ArrayListRegionInfo();
-1219long st, et;
-1220
-1221st = System.currentTimeMillis();
-1222for (RegionStateNode regionNode: 
regionStates.getRegionNodes()) {
-1223  if (regionNode.getState() == 
State.OPEN) {
-1224final ServerName serverName = 
regionNode.getRegionLocation();
-1225if 
(!master.getServerManager().isServerOnline(serverName)) {
-1226  
offlineServersWithOnlineRegions.add(serverName);
-1227}
-1228  } else if (regionNode.getState() 
== State.OFFLINE) {
-1229if 
(isTableEnabled(regionNode.getTable())) {
-1230  
regionsToAssign.add(regionNode.getRegionInfo());
-1231}
-1232  }
-1233}
-1234et = System.currentTimeMillis();
-1235LOG.info("[STEP-1] " + 
StringUtils.humanTimeDiff(et - st));
-1236
-1237// kill servers with online 
regions
-1238st = System.currentTimeMillis();
-1239for (ServerName serverName: 
offlineServersWithOnlineRegions) {
-1240  if 
(!master.getServerManager().isServerOnline(serverName)) {
-1241LOG.info("KILL RS hosting 
regions but not online " + serverName +
-1242  " (master=" + 
master.getServerName() + ")");
-1243killRegionServer(serverName);
-1244  }
-1245}
-1246et = System.currentTimeMillis();
-1247LOG.info("[STEP-2] " + 
StringUtils.humanTimeDiff(et - st));
-1248
-1249setFailoverCleanupDone(true);
-1250
-1251// assign offline regions

[39/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
index cfafcf8..90e8860 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":6,"i4":10,"i5":10,"i6":6,"i7":6,"i8":10,"i9":10,"i10":10,"i11":10,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":10,"i18":6,"i19":10,"i20":10,"i21":6,"i22":10,"i23":6,"i24":10,"i25":10,"i26":6};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":6,"i4":10,"i5":10,"i6":6,"i7":6,"i8":10,"i9":10,"i10":10,"i11":10,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":10,"i18":10,"i19":6,"i20":10,"i21":10,"i22":6,"i23":10,"i24":6,"i25":10,"i26":10,"i27":6};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -162,34 +162,38 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 Field and Description
 
 
+protected boolean
+atomic
+
+
 protected Durability
 durability
 
-
+
 protected http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Mapbyte[],http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListCell[]
 familyCellMaps
 
-
+
 protected int
 nextIndexToProcess
 
-
+
 protected HRegion.ObservedExceptionsInBatch
 observedExceptions
 
-
+
 protected T[]
 operations
 
-
+
 protected HRegion
 region
 
-
+
 protected OperationStatus[]
 retCodeDetails
 
-
+
 protected WALEdit[]
 walEditsFromCoprocessors
 
@@ -320,25 +324,29 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 getOrigLogSeqNum()
 
 
+(package private) boolean
+isAtomic()
+
+
 boolean
 isDone()
 
-
+
 abstract boolean
 isInReplay()
 
-
+
 boolean
 isOperationPending(intindex)
 
-
+
 MiniBatchOperationInProgressMutation
 lockRowsAndBuildMiniBatch(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegion.RowLockacquiredRowLocks)
 Creates Mini-batch of all operations [nextIndexToProcess, 
lastIndexExclusive) for which
  a row lock can be acquired.
 
 
-
+
 abstract void
 prepareMiniBatchOperations(MiniBatchOperationInProgressMutationminiBatchOp,
   longtimestamp,
@@ -347,15 +355,15 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
   count, tags and timestamp for all cells of all operations in a 
mini-batch.
 
 
-
+
 int
 size()
 
-
+
 abstract void
 startRegionOperation()
 
-
+
 void
 visitBatchOperations(booleanpendingOnly,
 intlastIndexExclusive,
@@ -363,12 +371,12 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 Helper method for visiting pending/ all batch 
operations
 
 
-
+
 protected void
 writeMiniBatchOperationsToMemStore(MiniBatchOperationInProgressMutationminiBatchOp,
   longwriteNumber)
 
-
+
 abstract MultiVersionConcurrencyControl.WriteEntry
 writeMiniBatchOperationsToMemStore(MiniBatchOperationInProgressMutationminiBatchOp,
   MultiVersionConcurrencyControl.WriteEntrywriteEntry)
@@ -463,12 +471,21 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 
-
+
 
 durability
 protectedDurability durability
 
 
+
+
+
+
+
+atomic
+protectedboolean atomic
+
+
 
 
 
@@ -485,7 +502,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 BatchOperation
-publicBatchOperation(HRegionregion,
+publicBatchOperation(HRegionregion,
   T[]operations)
 
 
@@ -503,7 +520,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 visitBatchOperations
-publicvoidvisitBatchOperations(booleanpendingOnly,
+publicvoidvisitBatchOperations(booleanpendingOnly,
  intlastIndexExclusive,
  HRegion.BatchOperation.Visitorvisitor)
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
@@ -520,7 +537,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getMutation
-public abstractMutationgetMutation(intindex)
+public abstractMutationgetMutation(intindex)
 
 
 
@@ -529,7 +546,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getNonceGroup
-public 

[35/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
index eaeba94..aae2da5 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html
@@ -105,7 +105,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-static interface RSRpcServices.LogDelegate
+static interface RSRpcServices.LogDelegate
 
 
 
@@ -151,7 +151,7 @@ var activeTableTab = "activeTableTab";
 
 
 logBatchWarning
-voidlogBatchWarning(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringfirstRegionName,
+voidlogBatchWarning(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringfirstRegionName,
  intsum,
  introwSizeWarnThreshold)
 



[34/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
index a2d5acd..4a9326a 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
@@ -516,12 +516,13 @@ implements 
 private void
-doBatchOp(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.RegionActionResult.Builderbuilder,
+doBatchOp(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.RegionActionResult.Builderbuilder,
  HRegionregion,
  OperationQuotaquota,
  http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Actionmutations,
  CellScannercells,
- ActivePolicyEnforcementspaceQuotaEnforcement)
+ ActivePolicyEnforcementspaceQuotaEnforcement,
+ booleanatomic)
 Execute a list of Put/Delete mutations.
 
 
@@ -1199,7 +1200,7 @@ implements 
 
 DEFAULT_LOG_DELEGATE
-private staticRSRpcServices.LogDelegate DEFAULT_LOG_DELEGATE
+private staticRSRpcServices.LogDelegate DEFAULT_LOG_DELEGATE
 
 
 
@@ -1208,7 +1209,7 @@ implements 
 
 ld
-private finalRSRpcServices.LogDelegate ld
+private finalRSRpcServices.LogDelegate ld
 
 
 
@@ -1218,7 +1219,7 @@ implements 
 SCANNER_ALREADY_CLOSED
 http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true;
 title="class or interface in java.lang">@Deprecated
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException SCANNER_ALREADY_CLOSED
+private static finalhttp://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException SCANNER_ALREADY_CLOSED
 Deprecated.
 
 
@@ -1236,7 +1237,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java
 
 
 RSRpcServices
-publicRSRpcServices(HRegionServerrs)
+publicRSRpcServices(HRegionServerrs)
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
 Throws:
@@ -1250,7 +1251,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java
 
 
 RSRpcServices
-RSRpcServices(HRegionServerrs,
+RSRpcServices(HRegionServerrs,
   RSRpcServices.LogDelegateld)
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
@@ -1396,7 +1397,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java
 
 
 checkAndRowMutate
-privatebooleancheckAndRowMutate(HRegionregion,
+privatebooleancheckAndRowMutate(HRegionregion,
   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Actionactions,
   CellScannercellScanner,
   byte[]row,
@@ -1411,7 +1412,6 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java
 
 Parameters:
 cellScanner - if non-null, the mutation data -- the Cell 
content.
-comparator - @throws IOException
 Throws:
 http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 
@@ -1423,7 +1423,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java
 
 
 append
-privateResultappend(HRegionregion,
+privateResultappend(HRegionregion,
   OperationQuotaquota,
   
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutationProtomutation,
   CellScannercellScanner,
@@ -1446,7 +1446,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java
 
 
 increment
-privateResultincrement(HRegionregion,
+privateResultincrement(HRegionregion,
  OperationQuotaquota,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutationProtomutation,
  CellScannercells,
@@ -1471,7 +1471,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java
 
 
 doNonAtomicRegionMutation
-privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListCellScannabledoNonAtomicRegionMutation(HRegionregion,
+privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 

[25/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
index 845f5e6..84e9e52 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html
@@ -98,241 +98,248 @@
 090  @Override
 091  public boolean visit(final Result 
r) throws IOException {
 092if (r !=  null  
!r.isEmpty()) {
-093  long st = 
System.currentTimeMillis();
-094  visitMetaEntry(visitor, r);
-095  long et = 
System.currentTimeMillis();
-096  LOG.info("[T] LOAD META PERF " 
+ StringUtils.humanTimeDiff(et - st));
-097} else if (isDebugEnabled) {
-098  LOG.debug("NULL result from 
meta - ignoring but this is strange.");
-099}
-100return true;
-101  }
-102});
-103  }
-104
-105  private void visitMetaEntry(final 
RegionStateVisitor visitor, final Result result)
-106  throws IOException {
-107final RegionLocations rl = 
MetaTableAccessor.getRegionLocations(result);
-108if (rl == null) return;
+093  long st = 0;
+094  if (LOG.isTraceEnabled()) {
+095st = 
System.currentTimeMillis();
+096  }
+097  visitMetaEntry(visitor, r);
+098  if (LOG.isTraceEnabled()) {
+099long et = 
System.currentTimeMillis();
+100LOG.trace("[T] LOAD META PERF 
" + StringUtils.humanTimeDiff(et - st));
+101  }
+102} else if (isDebugEnabled) {
+103  LOG.debug("NULL result from 
meta - ignoring but this is strange.");
+104}
+105return true;
+106  }
+107});
+108  }
 109
-110final HRegionLocation[] locations = 
rl.getRegionLocations();
-111if (locations == null) return;
-112
-113for (int i = 0; i  
locations.length; ++i) {
-114  final HRegionLocation hrl = 
locations[i];
-115  if (hrl == null) continue;
-116
-117  final RegionInfo regionInfo = 
hrl.getRegionInfo();
-118  if (regionInfo == null) continue;
-119
-120  final int replicaId = 
regionInfo.getReplicaId();
-121  final State state = 
getRegionState(result, replicaId);
-122
-123  final ServerName lastHost = 
hrl.getServerName();
-124  final ServerName regionLocation = 
getRegionServer(result, replicaId);
-125  final long openSeqNum = -1;
-126
-127  // TODO: move under trace, now is 
visible for debugging
-128  LOG.info(String.format("Load 
hbase:meta entry region=%s regionState=%s lastHost=%s regionLocation=%s",
-129regionInfo, state, lastHost, 
regionLocation));
-130
-131  
visitor.visitRegionState(regionInfo, state, regionLocation, lastHost, 
openSeqNum);
-132}
-133  }
-134
-135  public void updateRegionLocation(final 
RegionInfo regionInfo, final State state,
-136  final ServerName regionLocation, 
final ServerName lastHost, final long openSeqNum,
-137  final long pid)
-138  throws IOException {
-139if (regionInfo.isMetaRegion()) {
-140  updateMetaLocation(regionInfo, 
regionLocation);
-141} else {
-142  
updateUserRegionLocation(regionInfo, state, regionLocation, lastHost, 
openSeqNum, pid);
-143}
-144  }
-145
-146  public void updateRegionState(final 
long openSeqNum, final long pid,
-147  final RegionState newState, final 
RegionState oldState) throws IOException {
-148
updateRegionLocation(newState.getRegion(), newState.getState(), 
newState.getServerName(),
-149oldState != null ? 
oldState.getServerName() : null, openSeqNum, pid);
-150  }
-151
-152  protected void updateMetaLocation(final 
RegionInfo regionInfo, final ServerName serverName)
-153  throws IOException {
-154try {
-155  
MetaTableLocator.setMetaLocation(master.getZooKeeper(), serverName,
-156regionInfo.getReplicaId(), 
State.OPEN);
-157} catch (KeeperException e) {
-158  throw new IOException(e);
-159}
-160  }
-161
-162  protected void 
updateUserRegionLocation(final RegionInfo regionInfo, final State state,
-163  final ServerName regionLocation, 
final ServerName lastHost, final long openSeqNum,
-164  final long pid)
-165  throws IOException {
-166final int replicaId = 
regionInfo.getReplicaId();
-167final Put put = new 
Put(MetaTableAccessor.getMetaKeyForRegion(regionInfo));
-168MetaTableAccessor.addRegionInfo(put, 
regionInfo);
-169final StringBuilder info = new 
StringBuilder("pid=" + pid + " updating hbase:meta row=");
-170
info.append(regionInfo.getRegionNameAsString()).append(", 
regionState=").append(state);
-171if (openSeqNum = 0) {
-172  

[22/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html
index 8e0a6a2..08916af 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html
@@ -75,7 +75,7 @@
 067  };
 068
 069  protected static final State[] 
STATES_EXPECTED_ON_CLOSE = new State[] {
-070State.SPLITTING, State.SPLIT, // 
ServerCrashProcedure
+070State.SPLITTING, State.SPLIT, 
State.MERGING, // ServerCrashProcedure
 071State.OPEN,   // 
enabled/open
 072State.CLOSING // 
already in-progress (retrying)
 073  };
@@ -550,446 +550,454 @@
 542return 
!getTableRegionStates(tableName).isEmpty();
 543  }
 544
-545  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
-546return getRegionsOfTable(table, 
false);
-547  }
-548
-549  ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
-550final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
-551final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());
-552for (RegionStateNode node: nodes) {
-553  if (include(node, offline)) 
hris.add(node.getRegionInfo());
-554}
-555return hris;
-556  }
-557
-558  /**
-559   * Utility. Whether to include region 
in list of regions. Default is to
-560   * weed out split and offline 
regions.
-561   * @return True if we should include 
the codenode/code (do not include
-562   * if split or offline unless 
codeoffline/code is set to true.
-563   */
-564  boolean include(final RegionStateNode 
node, final boolean offline) {
-565if (LOG.isTraceEnabled()) {
-566  LOG.trace("WORKING ON " + node + " 
" + node.getRegionInfo());
-567}
-568if (node.isInState(State.SPLIT)) 
return false;
-569if (node.isInState(State.OFFLINE) 
 !offline) return false;
-570final RegionInfo hri = 
node.getRegionInfo();
-571return (!hri.isOffline()  
!hri.isSplit()) ||
-572((hri.isOffline() || 
hri.isSplit())  offline);
-573  }
-574
-575  /**
-576   * Returns the set of regions hosted by 
the specified server
-577   * @param serverName the server we are 
interested in
-578   * @return set of RegionInfo hosted by 
the specified server
-579   */
-580  public ListRegionInfo 
getServerRegionInfoSet(final ServerName serverName) {
-581final ServerStateNode serverInfo = 
getServerNode(serverName);
-582if (serverInfo == null) return 
Collections.emptyList();
-583
-584synchronized (serverInfo) {
-585  return 
serverInfo.getRegionInfoList();
-586}
-587  }
-588
-589  // 

-590  //  TODO: split helpers
-591  // 

-592  public void logSplit(final ServerName 
serverName) {
-593final ServerStateNode serverNode = 
getOrCreateServer(serverName);
-594synchronized (serverNode) {
-595  
serverNode.setState(ServerState.SPLITTING);
-596  /* THIS HAS TO BE WRONG. THIS IS 
SPLITTING OF REGION, NOT SPLITTING WALs.
-597  for (RegionStateNode regionNode: 
serverNode.getRegions()) {
-598synchronized (regionNode) {
-599  // TODO: Abort procedure if 
present
-600  
regionNode.setState(State.SPLITTING);
-601}
-602  }*/
-603}
-604  }
-605
-606  public void logSplit(final RegionInfo 
regionInfo) {
-607final RegionStateNode regionNode = 
getRegionNode(regionInfo);
-608synchronized (regionNode) {
-609  regionNode.setState(State.SPLIT);
-610}
-611  }
-612
-613  @VisibleForTesting
-614  public void updateRegionState(final 
RegionInfo regionInfo, final State state) {
-615final RegionStateNode regionNode = 
getOrCreateRegionNode(regionInfo);
+545  /**
+546   * @return Return online regions of 
table; does not include OFFLINE or SPLITTING regions.
+547   */
+548  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
+549return getRegionsOfTable(table, 
false);
+550  }
+551
+552  /**
+553   * @return Return the regions of the 
table; does not include OFFLINE unless you set
+554   * codeoffline/code to 
true. Does not include regions that are in the
+555   * {@link State#SPLIT} state.
+556   */
+557  public ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
+558final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
+559final ArrayListRegionInfo 
hris = new 

[51/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
Published site at .


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/a108018f
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/a108018f
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/a108018f

Branch: refs/heads/asf-site
Commit: a108018f14235b3904d9e73288bf8aabded259f3
Parents: d0ae29f
Author: jenkins 
Authored: Fri Nov 10 15:29:29 2017 +
Committer: jenkins 
Committed: Fri Nov 10 15:29:29 2017 +

--
 acid-semantics.html | 4 +-
 apache_hbase_reference_guide.pdf| 4 +-
 apidocs/allclasses-frame.html   | 1 +
 apidocs/allclasses-noframe.html | 1 +
 apidocs/index-all.html  |10 +
 apidocs/org/apache/hadoop/hbase/AuthUtil.html   | 4 +-
 .../apache/hadoop/hbase/CacheEvictionStats.html |   269 +
 .../hadoop/hbase/CallDroppedException.html  | 4 +-
 .../hbase/class-use/CacheEvictionStats.html |   169 +
 .../hadoop/hbase/class-use/TableName.html   | 2 +-
 .../org/apache/hadoop/hbase/client/Admin.html   | 6 +-
 .../org/apache/hadoop/hbase/package-frame.html  | 1 +
 .../apache/hadoop/hbase/package-summary.html|40 +-
 .../org/apache/hadoop/hbase/package-tree.html   | 1 +
 .../org/apache/hadoop/hbase/package-use.html|37 +-
 apidocs/overview-tree.html  | 1 +
 .../apache/hadoop/hbase/CacheEvictionStats.html |   126 +
 book.html   | 2 +-
 bulk-loads.html | 4 +-
 checkstyle-aggregate.html   |  1470 +-
 checkstyle.rss  |66 +-
 coc.html| 4 +-
 cygwin.html | 4 +-
 dependencies.html   | 4 +-
 dependency-convergence.html |10 +-
 dependency-info.html| 4 +-
 dependency-management.html  |78 +-
 devapidocs/allclasses-frame.html| 2 +
 devapidocs/allclasses-noframe.html  | 2 +
 devapidocs/constant-values.html |58 +-
 devapidocs/index-all.html   |74 +-
 .../org/apache/hadoop/hbase/Coprocessor.html| 2 +-
 .../hadoop/hbase/class-use/CellBuilder.html |22 +
 .../hadoop/hbase/class-use/CellScanner.html | 5 +-
 .../hadoop/hbase/class-use/Coprocessor.html |13 +
 .../hbase/class-use/CoprocessorEnvironment.html | 8 +
 .../hadoop/hbase/class-use/TableName.html   | 2 +-
 .../hadoop/hbase/client/class-use/Mutation.html |44 +-
 .../hbase/client/class-use/RegionInfo.html  | 2 +-
 .../hadoop/hbase/client/class-use/Scan.html |47 +-
 .../hadoop/hbase/client/package-tree.html   |24 +-
 .../apache/hadoop/hbase/client/package-use.html | 5 +
 .../hbase/coprocessor/RegionCoprocessor.html| 2 +-
 .../hbase/coprocessor/RegionObserver.html   | 2 +-
 .../coprocessor/class-use/ObserverContext.html  |36 +-
 .../class-use/RegionCoprocessor.html|13 +
 .../class-use/RegionCoprocessorEnvironment.html |36 +-
 .../coprocessor/class-use/RegionObserver.html   |21 +
 ...serverWithMetrics.ExampleRegionObserver.html |   119 +-
 .../ExampleRegionObserverWithMetrics.html   |42 +-
 .../coprocessor/example/RowCountEndpoint.html   | 4 +-
 .../example/ScanModifyingObserver.html  |   477 +
 .../example/ValueRewritingObserver.html |   515 +
 .../example/WriteHeavyIncrementObserver.html| 4 +-
 .../class-use/ScanModifyingObserver.html|   125 +
 .../class-use/ValueRewritingObserver.html   |   125 +
 .../coprocessor/example/package-frame.html  | 2 +
 .../coprocessor/example/package-summary.html|13 +
 .../hbase/coprocessor/example/package-tree.html | 2 +
 .../hadoop/hbase/executor/package-tree.html | 2 +-
 .../hadoop/hbase/filter/package-tree.html   |12 +-
 .../hadoop/hbase/io/hfile/package-tree.html | 6 +-
 .../apache/hadoop/hbase/ipc/package-tree.html   | 2 +-
 .../hadoop/hbase/mapreduce/package-tree.html| 6 +-
 .../master/assignment/AssignmentManager.html|   100 +-
 .../assignment/MergeTableRegionsProcedure.html  |   128 +-
 .../master/assignment/RegionStateStore.html |38 +-
 .../RegionStates.RegionFailedOpen.html  |22 +-
 .../hbase/master/assignment/RegionStates.html   |94 +-
 ...tTableRegionProcedure.StoreFileSplitter.html |12 +-
 .../assignment/SplitTableRegionProcedure.html   |82 +-
 .../class-use/RegionStates.ServerStateNode.html | 2 +-
 .../hbase/master/balancer/package-tree.html | 2 +-
 

[40/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
index 561d099..7d7a8b0 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
@@ -134,7 +134,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class SplitTableRegionProcedure
+public class SplitTableRegionProcedure
 extends AbstractStateMachineRegionProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionState
 The procedure to split a region in a table.
  Takes lock on the parent region.
@@ -516,7 +516,7 @@ extends 
 
 LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static finalorg.apache.commons.logging.Log LOG
 
 
 
@@ -525,7 +525,7 @@ extends 
 
 traceEnabled
-privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
+privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean traceEnabled
 
 
 
@@ -534,7 +534,7 @@ extends 
 
 daughter_1_RI
-privateRegionInfo daughter_1_RI
+privateRegionInfo daughter_1_RI
 
 
 
@@ -543,7 +543,7 @@ extends 
 
 daughter_2_RI
-privateRegionInfo daughter_2_RI
+privateRegionInfo daughter_2_RI
 
 
 
@@ -552,7 +552,7 @@ extends 
 
 bestSplitRow
-privatebyte[] bestSplitRow
+privatebyte[] bestSplitRow
 
 
 
@@ -561,7 +561,7 @@ extends 
 
 EXPECTED_SPLIT_STATES
-private staticRegionState.State[] EXPECTED_SPLIT_STATES
+private staticRegionState.State[] EXPECTED_SPLIT_STATES
 
 
 
@@ -578,7 +578,7 @@ extends 
 
 SplitTableRegionProcedure
-publicSplitTableRegionProcedure()
+publicSplitTableRegionProcedure()
 
 
 
@@ -587,7 +587,7 @@ extends 
 
 SplitTableRegionProcedure
-publicSplitTableRegionProcedure(MasterProcedureEnvenv,
+publicSplitTableRegionProcedure(MasterProcedureEnvenv,
  RegionInforegionToSplit,
  byte[]splitRow)
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
@@ -611,7 +611,7 @@ extends 
 
 checkSplittable
-privatevoidcheckSplittable(MasterProcedureEnvenv,
+privatevoidcheckSplittable(MasterProcedureEnvenv,
  RegionInforegionToSplit,
  byte[]splitRow)
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
@@ -632,7 +632,7 @@ extends 
 
 getDaughterRegionIdTimestamp
-private staticlonggetDaughterRegionIdTimestamp(RegionInfohri)
+private staticlonggetDaughterRegionIdTimestamp(RegionInfohri)
 Calculate daughter regionid to use.
 
 Parameters:
@@ -648,7 +648,7 @@ extends 
 
 executeFromState
-protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv,
+protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv,
   
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionStatestate)
throws http://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
 title="class or interface in java.lang">InterruptedException
 Description copied from 
class:StateMachineProcedure
@@ -671,7 +671,7 @@ extends 
 
 rollbackState
-protectedvoidrollbackState(MasterProcedureEnvenv,
+protectedvoidrollbackState(MasterProcedureEnvenv,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionStatestate)
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException,
  http://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
 title="class or interface in java.lang">InterruptedException
@@ -693,7 +693,7 @@ extends 
 
 isRollbackSupported
-protectedbooleanisRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionStatestate)
+protectedbooleanisRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionStatestate)
 Description copied from 
class:StateMachineProcedure
 Used by the default implementation of abort() to know if 
the current state can be aborted
  

[41/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
index e5363a7..0d3ffae 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public static final class RegionStates.RegionFailedOpen
+public static final class RegionStates.RegionFailedOpen
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 
 
@@ -229,7 +229,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 regionNode
-private finalRegionStates.RegionStateNode regionNode
+private finalRegionStates.RegionStateNode regionNode
 
 
 
@@ -238,7 +238,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 exception
-private volatilehttp://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exception exception
+private volatilehttp://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exception exception
 
 
 
@@ -247,7 +247,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 retries
-private volatileint retries
+private volatileint retries
 
 
 
@@ -264,7 +264,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 RegionFailedOpen
-publicRegionFailedOpen(RegionStates.RegionStateNoderegionNode)
+publicRegionFailedOpen(RegionStates.RegionStateNoderegionNode)
 
 
 
@@ -281,7 +281,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getRegionNode
-publicRegionStates.RegionStateNodegetRegionNode()
+publicRegionStates.RegionStateNodegetRegionNode()
 
 
 
@@ -290,7 +290,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getRegionInfo
-publicRegionInfogetRegionInfo()
+publicRegionInfogetRegionInfo()
 
 
 
@@ -299,7 +299,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 incrementAndGetRetries
-publicintincrementAndGetRetries()
+publicintincrementAndGetRetries()
 
 
 
@@ -308,7 +308,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getRetries
-publicintgetRetries()
+publicintgetRetries()
 
 
 
@@ -317,7 +317,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 setException
-publicvoidsetException(http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exceptionexception)
+publicvoidsetException(http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">Exceptionexception)
 
 
 
@@ -326,7 +326,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getException
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">ExceptiongetException()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">ExceptiongetException()
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html
index c71b9c3..e81ea36 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html
@@ -365,7 +365,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 getRegionsOfTable(TableNametable)
 
 
-(package private) http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListRegionInfo
 getRegionsOfTable(TableNametable,
  booleanoffline)
 
@@ -784,7 +784,11 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getRegionsOfTable
-publichttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 

[28/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
index f0775d0..40cd159 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
@@ -943,7 +943,7 @@
 935wakeServerReportEvent(serverNode);
 936  }
 937
-938  public void 
checkOnlineRegionsReportForMeta(final ServerStateNode serverNode,
+938  void 
checkOnlineRegionsReportForMeta(final ServerStateNode serverNode,
 939  final Setbyte[] 
regionNames) {
 940try {
 941  for (byte[] regionName: 
regionNames) {
@@ -959,7 +959,7 @@
 951final RegionStateNode regionNode 
= regionStates.getOrCreateRegionNode(hri);
 952LOG.info("META REPORTED: " + 
regionNode);
 953if (!reportTransition(regionNode, 
serverNode, TransitionCode.OPENED, 0)) {
-954  LOG.warn("META REPORTED but no 
procedure found");
+954  LOG.warn("META REPORTED but no 
procedure found (complete?)");
 955  
regionNode.setRegionLocation(serverNode.getServerName());
 956} else if (LOG.isTraceEnabled()) 
{
 957  LOG.trace("META REPORTED: " + 
regionNode);
@@ -1191,671 +1191,685 @@
 1183  public void visitRegionState(final 
RegionInfo regionInfo, final State state,
 1184  final ServerName 
regionLocation, final ServerName lastHost, final long openSeqNum) {
 1185final RegionStateNode regionNode 
= regionStates.getOrCreateRegionNode(regionInfo);
-1186synchronized (regionNode) {
-1187  if 
(!regionNode.isInTransition()) {
-1188
regionNode.setState(state);
-1189
regionNode.setLastHost(lastHost);
-1190
regionNode.setRegionLocation(regionLocation);
-1191
regionNode.setOpenSeqNum(openSeqNum);
-1192
-1193if (state == State.OPEN) {
-1194  assert regionLocation != 
null : "found null region location for " + regionNode;
-1195  
regionStates.addRegionToServer(regionLocation, regionNode);
-1196} else if (state == 
State.OFFLINE || regionInfo.isOffline()) {
-1197  
regionStates.addToOfflineRegions(regionNode);
-1198} else {
-1199  // These regions should 
have a procedure in replay
-1200  
regionStates.addRegionInTransition(regionNode, null);
-1201}
-1202  }
-1203}
-1204  }
-1205});
-1206
-1207// every assignment is blocked until 
meta is loaded.
-1208wakeMetaLoadedEvent();
-1209  }
-1210
-1211  // TODO: the assumption here is that 
if RSs are crashing while we are executing this
-1212  // they will be handled by the SSH 
that are put in the ServerManager "queue".
-1213  // we can integrate this a bit 
better.
-1214  private boolean 
processofflineServersWithOnlineRegions() {
-1215boolean failover = 
!master.getServerManager().getDeadServers().isEmpty();
-1216
-1217final SetServerName 
offlineServersWithOnlineRegions = new HashSetServerName();
-1218final ArrayListRegionInfo 
regionsToAssign = new ArrayListRegionInfo();
-1219long st, et;
-1220
-1221st = System.currentTimeMillis();
-1222for (RegionStateNode regionNode: 
regionStates.getRegionNodes()) {
-1223  if (regionNode.getState() == 
State.OPEN) {
-1224final ServerName serverName = 
regionNode.getRegionLocation();
-1225if 
(!master.getServerManager().isServerOnline(serverName)) {
-1226  
offlineServersWithOnlineRegions.add(serverName);
-1227}
-1228  } else if (regionNode.getState() 
== State.OFFLINE) {
-1229if 
(isTableEnabled(regionNode.getTable())) {
-1230  
regionsToAssign.add(regionNode.getRegionInfo());
-1231}
-1232  }
-1233}
-1234et = System.currentTimeMillis();
-1235LOG.info("[STEP-1] " + 
StringUtils.humanTimeDiff(et - st));
-1236
-1237// kill servers with online 
regions
-1238st = System.currentTimeMillis();
-1239for (ServerName serverName: 
offlineServersWithOnlineRegions) {
-1240  if 
(!master.getServerManager().isServerOnline(serverName)) {
-1241LOG.info("KILL RS hosting 
regions but not online " + serverName +
-1242  " (master=" + 
master.getServerName() + ")");
-1243killRegionServer(serverName);
-1244  }
-1245}
-1246et = System.currentTimeMillis();
-1247LOG.info("[STEP-2] " + 
StringUtils.humanTimeDiff(et - st));
-1248
-1249setFailoverCleanupDone(true);
-1250
-1251// assign offline regions
-1252st = System.currentTimeMillis();
-1253for (RegionInfo regionInfo: 
getOrderedRegions(regionsToAssign)) {

[16/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html
index e844201..496c92f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html
@@ -50,768 +50,774 @@
 042import 
org.apache.hadoop.hbase.HConstants;
 043import 
org.apache.hadoop.hbase.ServerName;
 044import 
org.apache.hadoop.hbase.TableName;
-045import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
-046import 
org.apache.hadoop.hbase.client.MasterSwitchType;
-047import 
org.apache.hadoop.hbase.client.Mutation;
-048import 
org.apache.hadoop.hbase.client.RegionInfo;
-049import 
org.apache.hadoop.hbase.client.RegionInfoBuilder;
-050import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-051import 
org.apache.hadoop.hbase.client.TableDescriptor;
-052import 
org.apache.hadoop.hbase.io.hfile.CacheConfig;
-053import 
org.apache.hadoop.hbase.master.MasterCoprocessorHost;
-054import 
org.apache.hadoop.hbase.master.MasterFileSystem;
-055import 
org.apache.hadoop.hbase.master.RegionState.State;
-056import 
org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode;
-057import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
-058import 
org.apache.hadoop.hbase.master.procedure.AbstractStateMachineRegionProcedure;
-059import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
-060import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
-061import 
org.apache.hadoop.hbase.procedure2.ProcedureMetrics;
-062import 
org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer;
-063import 
org.apache.hadoop.hbase.quotas.QuotaExceededException;
-064import 
org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
-065import 
org.apache.hadoop.hbase.regionserver.HStore;
-066import 
org.apache.hadoop.hbase.regionserver.HStoreFile;
-067import 
org.apache.hadoop.hbase.regionserver.StoreFileInfo;
-068import 
org.apache.hadoop.hbase.util.Bytes;
-069import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-070import 
org.apache.hadoop.hbase.util.FSUtils;
-071import 
org.apache.hadoop.hbase.util.Pair;
-072import 
org.apache.hadoop.hbase.util.Threads;
-073import 
org.apache.yetus.audience.InterfaceAudience;
-074
-075import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-076import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-077import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
-078import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
-079import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionState;
-080
-081/**
-082 * The procedure to split a region in a 
table.
-083 * Takes lock on the parent region.
-084 * It holds the lock for the life of the 
procedure.
-085 */
-086@InterfaceAudience.Private
-087public class SplitTableRegionProcedure
-088extends 
AbstractStateMachineRegionProcedureSplitTableRegionState {
-089  private static final Log LOG = 
LogFactory.getLog(SplitTableRegionProcedure.class);
-090  private Boolean traceEnabled = null;
-091  private RegionInfo daughter_1_RI;
-092  private RegionInfo daughter_2_RI;
-093  private byte[] bestSplitRow;
-094
-095  public SplitTableRegionProcedure() {
-096// Required by the Procedure 
framework to create the procedure on replay
-097  }
-098
-099  public SplitTableRegionProcedure(final 
MasterProcedureEnv env,
-100  final RegionInfo regionToSplit, 
final byte[] splitRow) throws IOException {
-101super(env, regionToSplit);
-102this.bestSplitRow = splitRow;
-103checkSplittable(env, regionToSplit, 
bestSplitRow);
-104final TableName table = 
regionToSplit.getTable();
-105final long rid = 
getDaughterRegionIdTimestamp(regionToSplit);
-106this.daughter_1_RI = 
RegionInfoBuilder.newBuilder(table)
-107
.setStartKey(regionToSplit.getStartKey())
-108.setEndKey(bestSplitRow)
-109.setSplit(false)
-110.setRegionId(rid)
-111.build();
-112this.daughter_2_RI = 
RegionInfoBuilder.newBuilder(table)
-113.setStartKey(bestSplitRow)
-114
.setEndKey(regionToSplit.getEndKey())
-115.setSplit(false)
-116.setRegionId(rid)
-117.build();
-118  }
-119
-120  /**
-121   * Check whether the region is 
splittable
-122   * @param env MasterProcedureEnv
-123   * @param regionToSplit parent Region 
to be split
-124   * @param 

[05/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, FSWALEntry#stampRegionSequenceId 

[31/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
index 7a70a5b..a7535b9 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html
@@ -43,99 +43,141 @@
 035import 
org.apache.hadoop.hbase.metrics.Counter;
 036import 
org.apache.hadoop.hbase.metrics.MetricRegistry;
 037import 
org.apache.hadoop.hbase.metrics.Timer;
-038
-039/**
-040 * An example coprocessor that collects 
some metrics to demonstrate the usage of exporting custom
-041 * metrics from the coprocessor.
-042 * p
-043 * These metrics will be available 
through the regular Hadoop metrics2 sinks (ganglia, opentsdb,
-044 * etc) as well as JMX output. You can 
view a snapshot of the metrics by going to the http web UI
-045 * of the regionserver page, something 
like http://myregionserverhost:16030/jmx
-046 * /p
-047 *
-048 * @see 
ExampleMasterObserverWithMetrics
-049 */
-050public class 
ExampleRegionObserverWithMetrics implements RegionCoprocessor {
-051
-052  private Counter preGetCounter;
-053  private Timer costlyOperationTimer;
-054  private ExampleRegionObserver 
observer;
-055
-056  class ExampleRegionObserver implements 
RegionCoprocessor, RegionObserver {
-057@Override
-058public OptionalRegionObserver 
getRegionObserver() {
-059  return Optional.of(this);
-060}
-061
-062@Override
-063public void 
preGetOp(ObserverContextRegionCoprocessorEnvironment e, Get get,
-064ListCell results) throws 
IOException {
-065  // Increment the Counter whenever 
the coprocessor is called
-066  preGetCounter.increment();
+038import 
org.apache.hadoop.hbase.regionserver.FlushLifeCycleTracker;
+039import 
org.apache.hadoop.hbase.regionserver.Store;
+040import 
org.apache.hadoop.hbase.regionserver.StoreFile;
+041import 
org.apache.hadoop.hbase.regionserver.compactions.CompactionLifeCycleTracker;
+042import 
org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
+043
+044/**
+045 * An example coprocessor that collects 
some metrics to demonstrate the usage of exporting custom
+046 * metrics from the coprocessor.
+047 * p
+048 * These metrics will be available 
through the regular Hadoop metrics2 sinks (ganglia, opentsdb,
+049 * etc) as well as JMX output. You can 
view a snapshot of the metrics by going to the http web UI
+050 * of the regionserver page, something 
like http://myregionserverhost:16030/jmx
+051 * /p
+052 *
+053 * @see 
ExampleMasterObserverWithMetrics
+054 */
+055public class 
ExampleRegionObserverWithMetrics implements RegionCoprocessor {
+056
+057  private Counter preGetCounter;
+058  private Counter flushCounter;
+059  private Counter 
filesCompactedCounter;
+060  private Timer costlyOperationTimer;
+061  private ExampleRegionObserver 
observer;
+062
+063  class ExampleRegionObserver implements 
RegionCoprocessor, RegionObserver {
+064@Override
+065public OptionalRegionObserver 
getRegionObserver() {
+066  return Optional.of(this);
 067}
 068
 069@Override
-070public void 
postGetOp(ObserverContextRegionCoprocessorEnvironment e, Get get,
+070public void 
preGetOp(ObserverContextRegionCoprocessorEnvironment e, Get get,
 071ListCell results) throws 
IOException {
-072  // do a costly (high latency) 
operation which we want to measure how long it takes by
-073  // using a Timer (which is a Meter 
and a Histogram).
-074  long start = System.nanoTime();
-075  try {
-076performCostlyOperation();
-077  } finally {
-078
costlyOperationTimer.updateNanos(System.nanoTime() - start);
-079  }
-080}
-081
-082private void performCostlyOperation() 
{
-083  try {
-084// simulate the operation by 
sleeping.
-085
Thread.sleep(ThreadLocalRandom.current().nextLong(100));
-086  } catch (InterruptedException 
ignore) {
-087  }
-088}
-089  }
-090
-091  @Override public 
OptionalRegionObserver getRegionObserver() {
-092return Optional.of(observer);
-093  }
-094
-095  @Override
-096  public void 
start(CoprocessorEnvironment env) throws IOException {
-097// start for the RegionServerObserver 
will be called only once in the lifetime of the
-098// server. We will construct and 
register all metrics that we will track across method
-099// invocations.
-100
-101if (env instanceof 
RegionCoprocessorEnvironment) {

[07/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, 

[17/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
index 8e0a6a2..08916af 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html
@@ -75,7 +75,7 @@
 067  };
 068
 069  protected static final State[] 
STATES_EXPECTED_ON_CLOSE = new State[] {
-070State.SPLITTING, State.SPLIT, // 
ServerCrashProcedure
+070State.SPLITTING, State.SPLIT, 
State.MERGING, // ServerCrashProcedure
 071State.OPEN,   // 
enabled/open
 072State.CLOSING // 
already in-progress (retrying)
 073  };
@@ -550,446 +550,454 @@
 542return 
!getTableRegionStates(tableName).isEmpty();
 543  }
 544
-545  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
-546return getRegionsOfTable(table, 
false);
-547  }
-548
-549  ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
-550final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
-551final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());
-552for (RegionStateNode node: nodes) {
-553  if (include(node, offline)) 
hris.add(node.getRegionInfo());
-554}
-555return hris;
-556  }
-557
-558  /**
-559   * Utility. Whether to include region 
in list of regions. Default is to
-560   * weed out split and offline 
regions.
-561   * @return True if we should include 
the codenode/code (do not include
-562   * if split or offline unless 
codeoffline/code is set to true.
-563   */
-564  boolean include(final RegionStateNode 
node, final boolean offline) {
-565if (LOG.isTraceEnabled()) {
-566  LOG.trace("WORKING ON " + node + " 
" + node.getRegionInfo());
-567}
-568if (node.isInState(State.SPLIT)) 
return false;
-569if (node.isInState(State.OFFLINE) 
 !offline) return false;
-570final RegionInfo hri = 
node.getRegionInfo();
-571return (!hri.isOffline()  
!hri.isSplit()) ||
-572((hri.isOffline() || 
hri.isSplit())  offline);
-573  }
-574
-575  /**
-576   * Returns the set of regions hosted by 
the specified server
-577   * @param serverName the server we are 
interested in
-578   * @return set of RegionInfo hosted by 
the specified server
-579   */
-580  public ListRegionInfo 
getServerRegionInfoSet(final ServerName serverName) {
-581final ServerStateNode serverInfo = 
getServerNode(serverName);
-582if (serverInfo == null) return 
Collections.emptyList();
-583
-584synchronized (serverInfo) {
-585  return 
serverInfo.getRegionInfoList();
-586}
-587  }
-588
-589  // 

-590  //  TODO: split helpers
-591  // 

-592  public void logSplit(final ServerName 
serverName) {
-593final ServerStateNode serverNode = 
getOrCreateServer(serverName);
-594synchronized (serverNode) {
-595  
serverNode.setState(ServerState.SPLITTING);
-596  /* THIS HAS TO BE WRONG. THIS IS 
SPLITTING OF REGION, NOT SPLITTING WALs.
-597  for (RegionStateNode regionNode: 
serverNode.getRegions()) {
-598synchronized (regionNode) {
-599  // TODO: Abort procedure if 
present
-600  
regionNode.setState(State.SPLITTING);
-601}
-602  }*/
-603}
-604  }
-605
-606  public void logSplit(final RegionInfo 
regionInfo) {
-607final RegionStateNode regionNode = 
getRegionNode(regionInfo);
-608synchronized (regionNode) {
-609  regionNode.setState(State.SPLIT);
-610}
-611  }
-612
-613  @VisibleForTesting
-614  public void updateRegionState(final 
RegionInfo regionInfo, final State state) {
-615final RegionStateNode regionNode = 
getOrCreateRegionNode(regionInfo);
+545  /**
+546   * @return Return online regions of 
table; does not include OFFLINE or SPLITTING regions.
+547   */
+548  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
+549return getRegionsOfTable(table, 
false);
+550  }
+551
+552  /**
+553   * @return Return the regions of the 
table; does not include OFFLINE unless you set
+554   * codeoffline/code to 
true. Does not include regions that are in the
+555   * {@link State#SPLIT} state.
+556   */
+557  public ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
+558final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
+559final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());
+560for (RegionStateNode node: nodes) {
+561  if 

[23/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
index 8e0a6a2..08916af 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
@@ -75,7 +75,7 @@
 067  };
 068
 069  protected static final State[] 
STATES_EXPECTED_ON_CLOSE = new State[] {
-070State.SPLITTING, State.SPLIT, // 
ServerCrashProcedure
+070State.SPLITTING, State.SPLIT, 
State.MERGING, // ServerCrashProcedure
 071State.OPEN,   // 
enabled/open
 072State.CLOSING // 
already in-progress (retrying)
 073  };
@@ -550,446 +550,454 @@
 542return 
!getTableRegionStates(tableName).isEmpty();
 543  }
 544
-545  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
-546return getRegionsOfTable(table, 
false);
-547  }
-548
-549  ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
-550final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
-551final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());
-552for (RegionStateNode node: nodes) {
-553  if (include(node, offline)) 
hris.add(node.getRegionInfo());
-554}
-555return hris;
-556  }
-557
-558  /**
-559   * Utility. Whether to include region 
in list of regions. Default is to
-560   * weed out split and offline 
regions.
-561   * @return True if we should include 
the codenode/code (do not include
-562   * if split or offline unless 
codeoffline/code is set to true.
-563   */
-564  boolean include(final RegionStateNode 
node, final boolean offline) {
-565if (LOG.isTraceEnabled()) {
-566  LOG.trace("WORKING ON " + node + " 
" + node.getRegionInfo());
-567}
-568if (node.isInState(State.SPLIT)) 
return false;
-569if (node.isInState(State.OFFLINE) 
 !offline) return false;
-570final RegionInfo hri = 
node.getRegionInfo();
-571return (!hri.isOffline()  
!hri.isSplit()) ||
-572((hri.isOffline() || 
hri.isSplit())  offline);
-573  }
-574
-575  /**
-576   * Returns the set of regions hosted by 
the specified server
-577   * @param serverName the server we are 
interested in
-578   * @return set of RegionInfo hosted by 
the specified server
-579   */
-580  public ListRegionInfo 
getServerRegionInfoSet(final ServerName serverName) {
-581final ServerStateNode serverInfo = 
getServerNode(serverName);
-582if (serverInfo == null) return 
Collections.emptyList();
-583
-584synchronized (serverInfo) {
-585  return 
serverInfo.getRegionInfoList();
-586}
-587  }
-588
-589  // 

-590  //  TODO: split helpers
-591  // 

-592  public void logSplit(final ServerName 
serverName) {
-593final ServerStateNode serverNode = 
getOrCreateServer(serverName);
-594synchronized (serverNode) {
-595  
serverNode.setState(ServerState.SPLITTING);
-596  /* THIS HAS TO BE WRONG. THIS IS 
SPLITTING OF REGION, NOT SPLITTING WALs.
-597  for (RegionStateNode regionNode: 
serverNode.getRegions()) {
-598synchronized (regionNode) {
-599  // TODO: Abort procedure if 
present
-600  
regionNode.setState(State.SPLITTING);
-601}
-602  }*/
-603}
-604  }
-605
-606  public void logSplit(final RegionInfo 
regionInfo) {
-607final RegionStateNode regionNode = 
getRegionNode(regionInfo);
-608synchronized (regionNode) {
-609  regionNode.setState(State.SPLIT);
-610}
-611  }
-612
-613  @VisibleForTesting
-614  public void updateRegionState(final 
RegionInfo regionInfo, final State state) {
-615final RegionStateNode regionNode = 
getOrCreateRegionNode(regionInfo);
+545  /**
+546   * @return Return online regions of 
table; does not include OFFLINE or SPLITTING regions.
+547   */
+548  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
+549return getRegionsOfTable(table, 
false);
+550  }
+551
+552  /**
+553   * @return Return the regions of the 
table; does not include OFFLINE unless you set
+554   * codeoffline/code to 
true. Does not include regions that are in the
+555   * {@link State#SPLIT} state.
+556   */
+557  public ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
+558final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
+559final ArrayListRegionInfo 
hris = new 

[18/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html
index 8e0a6a2..08916af 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html
@@ -75,7 +75,7 @@
 067  };
 068
 069  protected static final State[] 
STATES_EXPECTED_ON_CLOSE = new State[] {
-070State.SPLITTING, State.SPLIT, // 
ServerCrashProcedure
+070State.SPLITTING, State.SPLIT, 
State.MERGING, // ServerCrashProcedure
 071State.OPEN,   // 
enabled/open
 072State.CLOSING // 
already in-progress (retrying)
 073  };
@@ -550,446 +550,454 @@
 542return 
!getTableRegionStates(tableName).isEmpty();
 543  }
 544
-545  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
-546return getRegionsOfTable(table, 
false);
-547  }
-548
-549  ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
-550final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
-551final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());
-552for (RegionStateNode node: nodes) {
-553  if (include(node, offline)) 
hris.add(node.getRegionInfo());
-554}
-555return hris;
-556  }
-557
-558  /**
-559   * Utility. Whether to include region 
in list of regions. Default is to
-560   * weed out split and offline 
regions.
-561   * @return True if we should include 
the codenode/code (do not include
-562   * if split or offline unless 
codeoffline/code is set to true.
-563   */
-564  boolean include(final RegionStateNode 
node, final boolean offline) {
-565if (LOG.isTraceEnabled()) {
-566  LOG.trace("WORKING ON " + node + " 
" + node.getRegionInfo());
-567}
-568if (node.isInState(State.SPLIT)) 
return false;
-569if (node.isInState(State.OFFLINE) 
 !offline) return false;
-570final RegionInfo hri = 
node.getRegionInfo();
-571return (!hri.isOffline()  
!hri.isSplit()) ||
-572((hri.isOffline() || 
hri.isSplit())  offline);
-573  }
-574
-575  /**
-576   * Returns the set of regions hosted by 
the specified server
-577   * @param serverName the server we are 
interested in
-578   * @return set of RegionInfo hosted by 
the specified server
-579   */
-580  public ListRegionInfo 
getServerRegionInfoSet(final ServerName serverName) {
-581final ServerStateNode serverInfo = 
getServerNode(serverName);
-582if (serverInfo == null) return 
Collections.emptyList();
-583
-584synchronized (serverInfo) {
-585  return 
serverInfo.getRegionInfoList();
-586}
-587  }
-588
-589  // 

-590  //  TODO: split helpers
-591  // 

-592  public void logSplit(final ServerName 
serverName) {
-593final ServerStateNode serverNode = 
getOrCreateServer(serverName);
-594synchronized (serverNode) {
-595  
serverNode.setState(ServerState.SPLITTING);
-596  /* THIS HAS TO BE WRONG. THIS IS 
SPLITTING OF REGION, NOT SPLITTING WALs.
-597  for (RegionStateNode regionNode: 
serverNode.getRegions()) {
-598synchronized (regionNode) {
-599  // TODO: Abort procedure if 
present
-600  
regionNode.setState(State.SPLITTING);
-601}
-602  }*/
-603}
-604  }
-605
-606  public void logSplit(final RegionInfo 
regionInfo) {
-607final RegionStateNode regionNode = 
getRegionNode(regionInfo);
-608synchronized (regionNode) {
-609  regionNode.setState(State.SPLIT);
-610}
-611  }
-612
-613  @VisibleForTesting
-614  public void updateRegionState(final 
RegionInfo regionInfo, final State state) {
-615final RegionStateNode regionNode = 
getOrCreateRegionNode(regionInfo);
+545  /**
+546   * @return Return online regions of 
table; does not include OFFLINE or SPLITTING regions.
+547   */
+548  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
+549return getRegionsOfTable(table, 
false);
+550  }
+551
+552  /**
+553   * @return Return the regions of the 
table; does not include OFFLINE unless you set
+554   * codeoffline/code to 
true. Does not include regions that are in the
+555   * {@link State#SPLIT} state.
+556   */
+557  public ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
+558final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
+559final ArrayListRegionInfo 
hris = new 

[33/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html
index 73dd853..3dfd003 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html
@@ -176,6 +176,18 @@
 
 
 
+void
+ExampleRegionObserverWithMetrics.ExampleRegionObserver.postFlush(ObserverContextRegionCoprocessorEnvironmentc,
+ FlushLifeCycleTrackertracker)
+
+
+void
+ExampleRegionObserverWithMetrics.ExampleRegionObserver.postFlush(ObserverContextRegionCoprocessorEnvironmentc,
+ Storestore,
+ StoreFileresultFile,
+ FlushLifeCycleTrackertracker)
+
+
 InternalScanner
 WriteHeavyIncrementObserver.preFlush(ObserverContextRegionCoprocessorEnvironmentc,
 Storestore,

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html
index 5ba23fc..ad1c278 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html
@@ -103,13 +103,13 @@
 
 
 
-private static class
+(package private) static class
 HRegion.MutationBatchOperation
 Batch of mutation operations.
 
 
 
-private static class
+(package private) static class
 HRegion.ReplayBatchOperation
 Batch of mutations for replay.
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
index f872206..bfc586b 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html
@@ -674,12 +674,13 @@
 
 
 private void
-RSRpcServices.doBatchOp(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.RegionActionResult.Builderbuilder,
+RSRpcServices.doBatchOp(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.RegionActionResult.Builderbuilder,
  HRegionregion,
  OperationQuotaquota,
  http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in 
java.util">Listorg.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Actionmutations,
  CellScannercells,
- ActivePolicyEnforcementspaceQuotaEnforcement)
+ ActivePolicyEnforcementspaceQuotaEnforcement,
+ booleanatomic)
 Execute a list of Put/Delete mutations.
 
 
@@ -1051,8 +1052,9 @@
 MetricsRegionWrapperImpl(HRegionregion)
 
 
-MutationBatchOperation(HRegionregion,
+MutationBatchOperation(HRegionregion,
   Mutation[]operations,
+  booleanatomic,
   longnonceGroup,
   longnonce)
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html
index 2952f9f..d6acc81 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html
@@ -336,18 +336,27 @@
 
 
 InternalScanner
+ValueRewritingObserver.preCompact(ObserverContextRegionCoprocessorEnvironmentc,
+  Storestore,
+  InternalScannerscanner,
+  ScanTypescanType,
+  CompactionLifeCycleTrackertracker,
+  CompactionRequestrequest)
+
+
+InternalScanner
 WriteHeavyIncrementObserver.preFlush(ObserverContextRegionCoprocessorEnvironmentc,
 Storestore,
 InternalScannerscanner,
 FlushLifeCycleTrackertracker)
 
-
+
 InternalScanner
 WriteHeavyIncrementObserver.preMemStoreCompactionCompact(ObserverContextRegionCoprocessorEnvironmentc,
 Storestore,
 InternalScannerscanner)
 
-
+
 private 

[24/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
index 8e0a6a2..08916af 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
@@ -75,7 +75,7 @@
 067  };
 068
 069  protected static final State[] 
STATES_EXPECTED_ON_CLOSE = new State[] {
-070State.SPLITTING, State.SPLIT, // 
ServerCrashProcedure
+070State.SPLITTING, State.SPLIT, 
State.MERGING, // ServerCrashProcedure
 071State.OPEN,   // 
enabled/open
 072State.CLOSING // 
already in-progress (retrying)
 073  };
@@ -550,446 +550,454 @@
 542return 
!getTableRegionStates(tableName).isEmpty();
 543  }
 544
-545  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
-546return getRegionsOfTable(table, 
false);
-547  }
-548
-549  ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
-550final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
-551final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());
-552for (RegionStateNode node: nodes) {
-553  if (include(node, offline)) 
hris.add(node.getRegionInfo());
-554}
-555return hris;
-556  }
-557
-558  /**
-559   * Utility. Whether to include region 
in list of regions. Default is to
-560   * weed out split and offline 
regions.
-561   * @return True if we should include 
the codenode/code (do not include
-562   * if split or offline unless 
codeoffline/code is set to true.
-563   */
-564  boolean include(final RegionStateNode 
node, final boolean offline) {
-565if (LOG.isTraceEnabled()) {
-566  LOG.trace("WORKING ON " + node + " 
" + node.getRegionInfo());
-567}
-568if (node.isInState(State.SPLIT)) 
return false;
-569if (node.isInState(State.OFFLINE) 
 !offline) return false;
-570final RegionInfo hri = 
node.getRegionInfo();
-571return (!hri.isOffline()  
!hri.isSplit()) ||
-572((hri.isOffline() || 
hri.isSplit())  offline);
-573  }
-574
-575  /**
-576   * Returns the set of regions hosted by 
the specified server
-577   * @param serverName the server we are 
interested in
-578   * @return set of RegionInfo hosted by 
the specified server
-579   */
-580  public ListRegionInfo 
getServerRegionInfoSet(final ServerName serverName) {
-581final ServerStateNode serverInfo = 
getServerNode(serverName);
-582if (serverInfo == null) return 
Collections.emptyList();
-583
-584synchronized (serverInfo) {
-585  return 
serverInfo.getRegionInfoList();
-586}
-587  }
-588
-589  // 

-590  //  TODO: split helpers
-591  // 

-592  public void logSplit(final ServerName 
serverName) {
-593final ServerStateNode serverNode = 
getOrCreateServer(serverName);
-594synchronized (serverNode) {
-595  
serverNode.setState(ServerState.SPLITTING);
-596  /* THIS HAS TO BE WRONG. THIS IS 
SPLITTING OF REGION, NOT SPLITTING WALs.
-597  for (RegionStateNode regionNode: 
serverNode.getRegions()) {
-598synchronized (regionNode) {
-599  // TODO: Abort procedure if 
present
-600  
regionNode.setState(State.SPLITTING);
-601}
-602  }*/
-603}
-604  }
-605
-606  public void logSplit(final RegionInfo 
regionInfo) {
-607final RegionStateNode regionNode = 
getRegionNode(regionInfo);
-608synchronized (regionNode) {
-609  regionNode.setState(State.SPLIT);
-610}
-611  }
-612
-613  @VisibleForTesting
-614  public void updateRegionState(final 
RegionInfo regionInfo, final State state) {
-615final RegionStateNode regionNode = 
getOrCreateRegionNode(regionInfo);
+545  /**
+546   * @return Return online regions of 
table; does not include OFFLINE or SPLITTING regions.
+547   */
+548  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
+549return getRegionsOfTable(table, 
false);
+550  }
+551
+552  /**
+553   * @return Return the regions of the 
table; does not include OFFLINE unless you set
+554   * codeoffline/code to 
true. Does not include regions that are in the
+555   * {@link State#SPLIT} state.
+556   */
+557  public ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
+558final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
+559final 

[15/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
index e844201..496c92f 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html
@@ -50,768 +50,774 @@
 042import 
org.apache.hadoop.hbase.HConstants;
 043import 
org.apache.hadoop.hbase.ServerName;
 044import 
org.apache.hadoop.hbase.TableName;
-045import 
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
-046import 
org.apache.hadoop.hbase.client.MasterSwitchType;
-047import 
org.apache.hadoop.hbase.client.Mutation;
-048import 
org.apache.hadoop.hbase.client.RegionInfo;
-049import 
org.apache.hadoop.hbase.client.RegionInfoBuilder;
-050import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-051import 
org.apache.hadoop.hbase.client.TableDescriptor;
-052import 
org.apache.hadoop.hbase.io.hfile.CacheConfig;
-053import 
org.apache.hadoop.hbase.master.MasterCoprocessorHost;
-054import 
org.apache.hadoop.hbase.master.MasterFileSystem;
-055import 
org.apache.hadoop.hbase.master.RegionState.State;
-056import 
org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode;
-057import 
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
-058import 
org.apache.hadoop.hbase.master.procedure.AbstractStateMachineRegionProcedure;
-059import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
-060import 
org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
-061import 
org.apache.hadoop.hbase.procedure2.ProcedureMetrics;
-062import 
org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer;
-063import 
org.apache.hadoop.hbase.quotas.QuotaExceededException;
-064import 
org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
-065import 
org.apache.hadoop.hbase.regionserver.HStore;
-066import 
org.apache.hadoop.hbase.regionserver.HStoreFile;
-067import 
org.apache.hadoop.hbase.regionserver.StoreFileInfo;
-068import 
org.apache.hadoop.hbase.util.Bytes;
-069import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-070import 
org.apache.hadoop.hbase.util.FSUtils;
-071import 
org.apache.hadoop.hbase.util.Pair;
-072import 
org.apache.hadoop.hbase.util.Threads;
-073import 
org.apache.yetus.audience.InterfaceAudience;
-074
-075import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-076import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-077import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
-078import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
-079import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionState;
-080
-081/**
-082 * The procedure to split a region in a 
table.
-083 * Takes lock on the parent region.
-084 * It holds the lock for the life of the 
procedure.
-085 */
-086@InterfaceAudience.Private
-087public class SplitTableRegionProcedure
-088extends 
AbstractStateMachineRegionProcedureSplitTableRegionState {
-089  private static final Log LOG = 
LogFactory.getLog(SplitTableRegionProcedure.class);
-090  private Boolean traceEnabled = null;
-091  private RegionInfo daughter_1_RI;
-092  private RegionInfo daughter_2_RI;
-093  private byte[] bestSplitRow;
-094
-095  public SplitTableRegionProcedure() {
-096// Required by the Procedure 
framework to create the procedure on replay
-097  }
-098
-099  public SplitTableRegionProcedure(final 
MasterProcedureEnv env,
-100  final RegionInfo regionToSplit, 
final byte[] splitRow) throws IOException {
-101super(env, regionToSplit);
-102this.bestSplitRow = splitRow;
-103checkSplittable(env, regionToSplit, 
bestSplitRow);
-104final TableName table = 
regionToSplit.getTable();
-105final long rid = 
getDaughterRegionIdTimestamp(regionToSplit);
-106this.daughter_1_RI = 
RegionInfoBuilder.newBuilder(table)
-107
.setStartKey(regionToSplit.getStartKey())
-108.setEndKey(bestSplitRow)
-109.setSplit(false)
-110.setRegionId(rid)
-111.build();
-112this.daughter_2_RI = 
RegionInfoBuilder.newBuilder(table)
-113.setStartKey(bestSplitRow)
-114
.setEndKey(regionToSplit.getEndKey())
-115.setSplit(false)
-116.setRegionId(rid)
-117.build();
-118  }
-119
-120  /**
-121   * Check whether the region is 
splittable
-122   * @param env MasterProcedureEnv
-123   * @param regionToSplit parent Region 
to be split
-124   * @param splitRow if splitRow is not 
specified, will first try to get bestSplitRow from RS
-125   * 

[29/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
index f0775d0..40cd159 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html
@@ -943,7 +943,7 @@
 935wakeServerReportEvent(serverNode);
 936  }
 937
-938  public void 
checkOnlineRegionsReportForMeta(final ServerStateNode serverNode,
+938  void 
checkOnlineRegionsReportForMeta(final ServerStateNode serverNode,
 939  final Setbyte[] 
regionNames) {
 940try {
 941  for (byte[] regionName: 
regionNames) {
@@ -959,7 +959,7 @@
 951final RegionStateNode regionNode 
= regionStates.getOrCreateRegionNode(hri);
 952LOG.info("META REPORTED: " + 
regionNode);
 953if (!reportTransition(regionNode, 
serverNode, TransitionCode.OPENED, 0)) {
-954  LOG.warn("META REPORTED but no 
procedure found");
+954  LOG.warn("META REPORTED but no 
procedure found (complete?)");
 955  
regionNode.setRegionLocation(serverNode.getServerName());
 956} else if (LOG.isTraceEnabled()) 
{
 957  LOG.trace("META REPORTED: " + 
regionNode);
@@ -1191,671 +1191,685 @@
 1183  public void visitRegionState(final 
RegionInfo regionInfo, final State state,
 1184  final ServerName 
regionLocation, final ServerName lastHost, final long openSeqNum) {
 1185final RegionStateNode regionNode 
= regionStates.getOrCreateRegionNode(regionInfo);
-1186synchronized (regionNode) {
-1187  if 
(!regionNode.isInTransition()) {
-1188
regionNode.setState(state);
-1189
regionNode.setLastHost(lastHost);
-1190
regionNode.setRegionLocation(regionLocation);
-1191
regionNode.setOpenSeqNum(openSeqNum);
-1192
-1193if (state == State.OPEN) {
-1194  assert regionLocation != 
null : "found null region location for " + regionNode;
-1195  
regionStates.addRegionToServer(regionLocation, regionNode);
-1196} else if (state == 
State.OFFLINE || regionInfo.isOffline()) {
-1197  
regionStates.addToOfflineRegions(regionNode);
-1198} else {
-1199  // These regions should 
have a procedure in replay
-1200  
regionStates.addRegionInTransition(regionNode, null);
-1201}
-1202  }
-1203}
-1204  }
-1205});
-1206
-1207// every assignment is blocked until 
meta is loaded.
-1208wakeMetaLoadedEvent();
-1209  }
-1210
-1211  // TODO: the assumption here is that 
if RSs are crashing while we are executing this
-1212  // they will be handled by the SSH 
that are put in the ServerManager "queue".
-1213  // we can integrate this a bit 
better.
-1214  private boolean 
processofflineServersWithOnlineRegions() {
-1215boolean failover = 
!master.getServerManager().getDeadServers().isEmpty();
-1216
-1217final SetServerName 
offlineServersWithOnlineRegions = new HashSetServerName();
-1218final ArrayListRegionInfo 
regionsToAssign = new ArrayListRegionInfo();
-1219long st, et;
-1220
-1221st = System.currentTimeMillis();
-1222for (RegionStateNode regionNode: 
regionStates.getRegionNodes()) {
-1223  if (regionNode.getState() == 
State.OPEN) {
-1224final ServerName serverName = 
regionNode.getRegionLocation();
-1225if 
(!master.getServerManager().isServerOnline(serverName)) {
-1226  
offlineServersWithOnlineRegions.add(serverName);
-1227}
-1228  } else if (regionNode.getState() 
== State.OFFLINE) {
-1229if 
(isTableEnabled(regionNode.getTable())) {
-1230  
regionsToAssign.add(regionNode.getRegionInfo());
-1231}
-1232  }
-1233}
-1234et = System.currentTimeMillis();
-1235LOG.info("[STEP-1] " + 
StringUtils.humanTimeDiff(et - st));
-1236
-1237// kill servers with online 
regions
-1238st = System.currentTimeMillis();
-1239for (ServerName serverName: 
offlineServersWithOnlineRegions) {
-1240  if 
(!master.getServerManager().isServerOnline(serverName)) {
-1241LOG.info("KILL RS hosting 
regions but not online " + serverName +
-1242  " (master=" + 
master.getServerName() + ")");
-1243killRegionServer(serverName);
-1244  }
-1245}
-1246et = System.currentTimeMillis();
-1247LOG.info("[STEP-2] " + 
StringUtils.humanTimeDiff(et - st));
-1248
-1249setFailoverCleanupDone(true);
-1250
-1251// assign offline regions
-1252   

[06/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, 

[20/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html
index 8e0a6a2..08916af 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html
@@ -75,7 +75,7 @@
 067  };
 068
 069  protected static final State[] 
STATES_EXPECTED_ON_CLOSE = new State[] {
-070State.SPLITTING, State.SPLIT, // 
ServerCrashProcedure
+070State.SPLITTING, State.SPLIT, 
State.MERGING, // ServerCrashProcedure
 071State.OPEN,   // 
enabled/open
 072State.CLOSING // 
already in-progress (retrying)
 073  };
@@ -550,446 +550,454 @@
 542return 
!getTableRegionStates(tableName).isEmpty();
 543  }
 544
-545  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
-546return getRegionsOfTable(table, 
false);
-547  }
-548
-549  ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
-550final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
-551final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());
-552for (RegionStateNode node: nodes) {
-553  if (include(node, offline)) 
hris.add(node.getRegionInfo());
-554}
-555return hris;
-556  }
-557
-558  /**
-559   * Utility. Whether to include region 
in list of regions. Default is to
-560   * weed out split and offline 
regions.
-561   * @return True if we should include 
the codenode/code (do not include
-562   * if split or offline unless 
codeoffline/code is set to true.
-563   */
-564  boolean include(final RegionStateNode 
node, final boolean offline) {
-565if (LOG.isTraceEnabled()) {
-566  LOG.trace("WORKING ON " + node + " 
" + node.getRegionInfo());
-567}
-568if (node.isInState(State.SPLIT)) 
return false;
-569if (node.isInState(State.OFFLINE) 
 !offline) return false;
-570final RegionInfo hri = 
node.getRegionInfo();
-571return (!hri.isOffline()  
!hri.isSplit()) ||
-572((hri.isOffline() || 
hri.isSplit())  offline);
-573  }
-574
-575  /**
-576   * Returns the set of regions hosted by 
the specified server
-577   * @param serverName the server we are 
interested in
-578   * @return set of RegionInfo hosted by 
the specified server
-579   */
-580  public ListRegionInfo 
getServerRegionInfoSet(final ServerName serverName) {
-581final ServerStateNode serverInfo = 
getServerNode(serverName);
-582if (serverInfo == null) return 
Collections.emptyList();
-583
-584synchronized (serverInfo) {
-585  return 
serverInfo.getRegionInfoList();
-586}
-587  }
-588
-589  // 

-590  //  TODO: split helpers
-591  // 

-592  public void logSplit(final ServerName 
serverName) {
-593final ServerStateNode serverNode = 
getOrCreateServer(serverName);
-594synchronized (serverNode) {
-595  
serverNode.setState(ServerState.SPLITTING);
-596  /* THIS HAS TO BE WRONG. THIS IS 
SPLITTING OF REGION, NOT SPLITTING WALs.
-597  for (RegionStateNode regionNode: 
serverNode.getRegions()) {
-598synchronized (regionNode) {
-599  // TODO: Abort procedure if 
present
-600  
regionNode.setState(State.SPLITTING);
-601}
-602  }*/
-603}
-604  }
-605
-606  public void logSplit(final RegionInfo 
regionInfo) {
-607final RegionStateNode regionNode = 
getRegionNode(regionInfo);
-608synchronized (regionNode) {
-609  regionNode.setState(State.SPLIT);
-610}
-611  }
-612
-613  @VisibleForTesting
-614  public void updateRegionState(final 
RegionInfo regionInfo, final State state) {
-615final RegionStateNode regionNode = 
getOrCreateRegionNode(regionInfo);
+545  /**
+546   * @return Return online regions of 
table; does not include OFFLINE or SPLITTING regions.
+547   */
+548  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
+549return getRegionsOfTable(table, 
false);
+550  }
+551
+552  /**
+553   * @return Return the regions of the 
table; does not include OFFLINE unless you set
+554   * codeoffline/code to 
true. Does not include regions that are in the
+555   * {@link State#SPLIT} state.
+556   */
+557  public ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
+558final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
+559final ArrayListRegionInfo 
hris = new 

[03/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, 

[01/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site d0ae29f24 -> a108018f1


http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is 

[13/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, 

[02/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, FSWALEntry#stampRegionSequenceId won't stamp sequence 

[26/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html
index 845f5e6..84e9e52 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html
@@ -98,241 +98,248 @@
 090  @Override
 091  public boolean visit(final Result 
r) throws IOException {
 092if (r !=  null  
!r.isEmpty()) {
-093  long st = 
System.currentTimeMillis();
-094  visitMetaEntry(visitor, r);
-095  long et = 
System.currentTimeMillis();
-096  LOG.info("[T] LOAD META PERF " 
+ StringUtils.humanTimeDiff(et - st));
-097} else if (isDebugEnabled) {
-098  LOG.debug("NULL result from 
meta - ignoring but this is strange.");
-099}
-100return true;
-101  }
-102});
-103  }
-104
-105  private void visitMetaEntry(final 
RegionStateVisitor visitor, final Result result)
-106  throws IOException {
-107final RegionLocations rl = 
MetaTableAccessor.getRegionLocations(result);
-108if (rl == null) return;
+093  long st = 0;
+094  if (LOG.isTraceEnabled()) {
+095st = 
System.currentTimeMillis();
+096  }
+097  visitMetaEntry(visitor, r);
+098  if (LOG.isTraceEnabled()) {
+099long et = 
System.currentTimeMillis();
+100LOG.trace("[T] LOAD META PERF 
" + StringUtils.humanTimeDiff(et - st));
+101  }
+102} else if (isDebugEnabled) {
+103  LOG.debug("NULL result from 
meta - ignoring but this is strange.");
+104}
+105return true;
+106  }
+107});
+108  }
 109
-110final HRegionLocation[] locations = 
rl.getRegionLocations();
-111if (locations == null) return;
-112
-113for (int i = 0; i  
locations.length; ++i) {
-114  final HRegionLocation hrl = 
locations[i];
-115  if (hrl == null) continue;
-116
-117  final RegionInfo regionInfo = 
hrl.getRegionInfo();
-118  if (regionInfo == null) continue;
-119
-120  final int replicaId = 
regionInfo.getReplicaId();
-121  final State state = 
getRegionState(result, replicaId);
-122
-123  final ServerName lastHost = 
hrl.getServerName();
-124  final ServerName regionLocation = 
getRegionServer(result, replicaId);
-125  final long openSeqNum = -1;
-126
-127  // TODO: move under trace, now is 
visible for debugging
-128  LOG.info(String.format("Load 
hbase:meta entry region=%s regionState=%s lastHost=%s regionLocation=%s",
-129regionInfo, state, lastHost, 
regionLocation));
-130
-131  
visitor.visitRegionState(regionInfo, state, regionLocation, lastHost, 
openSeqNum);
-132}
-133  }
-134
-135  public void updateRegionLocation(final 
RegionInfo regionInfo, final State state,
-136  final ServerName regionLocation, 
final ServerName lastHost, final long openSeqNum,
-137  final long pid)
-138  throws IOException {
-139if (regionInfo.isMetaRegion()) {
-140  updateMetaLocation(regionInfo, 
regionLocation);
-141} else {
-142  
updateUserRegionLocation(regionInfo, state, regionLocation, lastHost, 
openSeqNum, pid);
-143}
-144  }
-145
-146  public void updateRegionState(final 
long openSeqNum, final long pid,
-147  final RegionState newState, final 
RegionState oldState) throws IOException {
-148
updateRegionLocation(newState.getRegion(), newState.getState(), 
newState.getServerName(),
-149oldState != null ? 
oldState.getServerName() : null, openSeqNum, pid);
-150  }
-151
-152  protected void updateMetaLocation(final 
RegionInfo regionInfo, final ServerName serverName)
-153  throws IOException {
-154try {
-155  
MetaTableLocator.setMetaLocation(master.getZooKeeper(), serverName,
-156regionInfo.getReplicaId(), 
State.OPEN);
-157} catch (KeeperException e) {
-158  throw new IOException(e);
-159}
-160  }
-161
-162  protected void 
updateUserRegionLocation(final RegionInfo regionInfo, final State state,
-163  final ServerName regionLocation, 
final ServerName lastHost, final long openSeqNum,
-164  final long pid)
-165  throws IOException {
-166final int replicaId = 
regionInfo.getReplicaId();
-167final Put put = new 
Put(MetaTableAccessor.getMetaKeyForRegion(regionInfo));
-168MetaTableAccessor.addRegionInfo(put, 
regionInfo);
-169final StringBuilder info = new 
StringBuilder("pid=" + pid + " updating hbase:meta row=");
-170
info.append(regionInfo.getRegionNameAsString()).append(", 

[04/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, FSWALEntry#stampRegionSequenceId won't 

[08/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, FSWALEntry#stampRegionSequenceId won't stamp 

[11/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, FSWALEntry#stampRegionSequenceId won't stamp 

[14/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html
index ffb1464..da77107 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html
@@ -6,7 +6,7 @@
 
 
 
-001/**
+001/*
 002 * Licensed to the Apache Software 
Foundation (ASF) under one
 003 * or more contributor license 
agreements.  See the NOTICE file
 004 * distributed with this work for 
additional information
@@ -128,9 +128,9 @@
 120// Get the replica count
 121int regionReplicaCount = 
hTableDescriptor.getRegionReplication();
 122
-123// Get the regions for the table 
from the memory
+123// Get the regions for the table 
from memory; get both online and offline regions ('true').
 124ListRegionInfo 
regionsOfTable =
-125
env.getAssignmentManager().getRegionStates().getRegionsOfTable(tableName);
+125
env.getAssignmentManager().getRegionStates().getRegionsOfTable(tableName, 
true);
 126
 127if (regionReplicaCount  1) 
{
 128  int currentMaxReplica = 0;



[10/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, FSWALEntry#stampRegionSequenceId 

[12/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, FSWALEntry#stampRegionSequenceId won't stamp sequence 

[21/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html
index 8e0a6a2..08916af 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html
@@ -75,7 +75,7 @@
 067  };
 068
 069  protected static final State[] 
STATES_EXPECTED_ON_CLOSE = new State[] {
-070State.SPLITTING, State.SPLIT, // 
ServerCrashProcedure
+070State.SPLITTING, State.SPLIT, 
State.MERGING, // ServerCrashProcedure
 071State.OPEN,   // 
enabled/open
 072State.CLOSING // 
already in-progress (retrying)
 073  };
@@ -550,446 +550,454 @@
 542return 
!getTableRegionStates(tableName).isEmpty();
 543  }
 544
-545  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
-546return getRegionsOfTable(table, 
false);
-547  }
-548
-549  ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
-550final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
-551final ArrayListRegionInfo 
hris = new ArrayListRegionInfo(nodes.size());
-552for (RegionStateNode node: nodes) {
-553  if (include(node, offline)) 
hris.add(node.getRegionInfo());
-554}
-555return hris;
-556  }
-557
-558  /**
-559   * Utility. Whether to include region 
in list of regions. Default is to
-560   * weed out split and offline 
regions.
-561   * @return True if we should include 
the codenode/code (do not include
-562   * if split or offline unless 
codeoffline/code is set to true.
-563   */
-564  boolean include(final RegionStateNode 
node, final boolean offline) {
-565if (LOG.isTraceEnabled()) {
-566  LOG.trace("WORKING ON " + node + " 
" + node.getRegionInfo());
-567}
-568if (node.isInState(State.SPLIT)) 
return false;
-569if (node.isInState(State.OFFLINE) 
 !offline) return false;
-570final RegionInfo hri = 
node.getRegionInfo();
-571return (!hri.isOffline()  
!hri.isSplit()) ||
-572((hri.isOffline() || 
hri.isSplit())  offline);
-573  }
-574
-575  /**
-576   * Returns the set of regions hosted by 
the specified server
-577   * @param serverName the server we are 
interested in
-578   * @return set of RegionInfo hosted by 
the specified server
-579   */
-580  public ListRegionInfo 
getServerRegionInfoSet(final ServerName serverName) {
-581final ServerStateNode serverInfo = 
getServerNode(serverName);
-582if (serverInfo == null) return 
Collections.emptyList();
-583
-584synchronized (serverInfo) {
-585  return 
serverInfo.getRegionInfoList();
-586}
-587  }
-588
-589  // 

-590  //  TODO: split helpers
-591  // 

-592  public void logSplit(final ServerName 
serverName) {
-593final ServerStateNode serverNode = 
getOrCreateServer(serverName);
-594synchronized (serverNode) {
-595  
serverNode.setState(ServerState.SPLITTING);
-596  /* THIS HAS TO BE WRONG. THIS IS 
SPLITTING OF REGION, NOT SPLITTING WALs.
-597  for (RegionStateNode regionNode: 
serverNode.getRegions()) {
-598synchronized (regionNode) {
-599  // TODO: Abort procedure if 
present
-600  
regionNode.setState(State.SPLITTING);
-601}
-602  }*/
-603}
-604  }
-605
-606  public void logSplit(final RegionInfo 
regionInfo) {
-607final RegionStateNode regionNode = 
getRegionNode(regionInfo);
-608synchronized (regionNode) {
-609  regionNode.setState(State.SPLIT);
-610}
-611  }
-612
-613  @VisibleForTesting
-614  public void updateRegionState(final 
RegionInfo regionInfo, final State state) {
-615final RegionStateNode regionNode = 
getOrCreateRegionNode(regionInfo);
+545  /**
+546   * @return Return online regions of 
table; does not include OFFLINE or SPLITTING regions.
+547   */
+548  public ListRegionInfo 
getRegionsOfTable(final TableName table) {
+549return getRegionsOfTable(table, 
false);
+550  }
+551
+552  /**
+553   * @return Return the regions of the 
table; does not include OFFLINE unless you set
+554   * codeoffline/code to 
true. Does not include regions that are in the
+555   * {@link State#SPLIT} state.
+556   */
+557  public ListRegionInfo 
getRegionsOfTable(final TableName table, final boolean offline) {
+558final 
ArrayListRegionStateNode nodes = getTableRegionStateNodes(table);
+559

[09/51] [partial] hbase-site git commit: Published site at .

2017-11-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
index 5c5e0eb..7b41db2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html
@@ -2960,5416 +2960,5447 @@
 2952protected final 
ObservedExceptionsInBatch observedExceptions;
 2953//Durability of the batch (highest 
durability of all operations)
 2954protected Durability durability;
-2955
-2956public BatchOperation(final HRegion 
region, T[] operations) {
-2957  this.operations = operations;
-2958  this.retCodeDetails = new 
OperationStatus[operations.length];
-2959  Arrays.fill(this.retCodeDetails, 
OperationStatus.NOT_RUN);
-2960  this.walEditsFromCoprocessors = 
new WALEdit[operations.length];
-2961  familyCellMaps = new 
Map[operations.length];
-2962
-2963  this.region = region;
-2964  observedExceptions = new 
ObservedExceptionsInBatch();
-2965  durability = 
Durability.USE_DEFAULT;
-2966}
-2967
-2968/**
-2969 * Visitor interface for batch 
operations
-2970 */
-2971@FunctionalInterface
-2972public interface Visitor {
-2973  /**
-2974   * @param index operation index
-2975   * @return If true continue 
visiting remaining entries, break otherwise
-2976   */
-2977  boolean visit(int index) throws 
IOException;
-2978}
-2979
-2980/**
-2981 * Helper method for visiting 
pending/ all batch operations
-2982 */
-2983public void 
visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor 
visitor)
-2984throws IOException {
-2985  assert lastIndexExclusive = 
this.size();
-2986  for (int i = nextIndexToProcess; i 
 lastIndexExclusive; i++) {
-2987if (!pendingOnly || 
isOperationPending(i)) {
-2988  if (!visitor.visit(i)) {
-2989break;
-2990  }
-2991}
-2992  }
-2993}
-2994
-2995public abstract Mutation 
getMutation(int index);
-2996public abstract long 
getNonceGroup(int index);
-2997public abstract long getNonce(int 
index);
-2998/** This method is potentially 
expensive and useful mostly for non-replay CP path. */
-2999public abstract Mutation[] 
getMutationsForCoprocs();
-3000public abstract boolean 
isInReplay();
-3001public abstract long 
getOrigLogSeqNum();
-3002public abstract void 
startRegionOperation() throws IOException;
-3003public abstract void 
closeRegionOperation() throws IOException;
-3004
-3005/**
-3006 * Validates each mutation and 
prepares a batch for write. If necessary (non-replay case), runs
-3007 * CP prePut()/ preDelete() hooks 
for all mutations in a batch. This is intended to operate on
-3008 * entire batch and will be called 
from outside of class to check and prepare batch. This can
-3009 * be implemented by calling helper 
method {@link #checkAndPrepareMutation(int, long)} in a
-3010 * 'for' loop over mutations.
-3011 */
-3012public abstract void 
checkAndPrepare() throws IOException;
-3013
-3014/**
-3015 * Implement any Put request 
specific check and prepare logic here. Please refer to
-3016 * {@link 
#checkAndPrepareMutation(Mutation, long)} for how its used.
-3017 */
-3018protected abstract void 
checkAndPreparePut(final Put p) throws IOException;
-3019
-3020/**
-3021 *  If necessary, calls 
preBatchMutate() CP hook for a mini-batch and updates metrics, cell
-3022 *  count, tags and timestamp for 
all cells of all operations in a mini-batch.
-3023 */
-3024public abstract void 
prepareMiniBatchOperations(MiniBatchOperationInProgressMutation
-3025miniBatchOp, long timestamp, 
final ListRowLock acquiredRowLocks) throws IOException;
-3026
-3027/**
-3028 * Write mini-batch operations to 
MemStore
-3029 */
-3030public abstract WriteEntry 
writeMiniBatchOperationsToMemStore(
-3031final 
MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry 
writeEntry)
-3032throws IOException;
-3033
-3034protected void 
writeMiniBatchOperationsToMemStore(
-3035final 
MiniBatchOperationInProgressMutation miniBatchOp, final long 
writeNumber)
-3036throws IOException {
-3037  MemStoreSizing memStoreAccounting 
= new MemStoreSizing();
-3038  visitBatchOperations(true, 
miniBatchOp.getLastIndexExclusive(), (int index) - {
-3039// We need to update the 
sequence id for following reasons.
-3040// 1) If the op is in replay 
mode, FSWALEntry#stampRegionSequenceId won't stamp sequence id.
-3041

hbase git commit: HBASE-16459 Remove unused hbase shell --format option

2017-11-10 Thread appy
Repository: hbase
Updated Branches:
  refs/heads/master 35a1e4334 -> 31234eb86


HBASE-16459 Remove unused hbase shell --format option


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/31234eb8
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/31234eb8
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/31234eb8

Branch: refs/heads/master
Commit: 31234eb862fdd7ee4917a2f74001182565ffbfd9
Parents: 35a1e43
Author: Apekshit Sharma 
Authored: Thu Nov 9 15:03:13 2017 -0800
Committer: Apekshit Sharma 
Committed: Fri Nov 10 00:08:02 2017 -0800

--
 bin/hirb.rb | 17 +
 1 file changed, 1 insertion(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/31234eb8/bin/hirb.rb
--
diff --git a/bin/hirb.rb b/bin/hirb.rb
index c06c83b..35895f8 100644
--- a/bin/hirb.rb
+++ b/bin/hirb.rb
@@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources)
 cmdline_help = <