hbase git commit: HBASE-19187 Remove option to create on heap bucket cache.
Repository: hbase Updated Branches: refs/heads/branch-2 30f55f231 -> 3ee76f857 HBASE-19187 Remove option to create on heap bucket cache. Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/3ee76f85 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/3ee76f85 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/3ee76f85 Branch: refs/heads/branch-2 Commit: 3ee76f8573387097c896d0611c30504ed7e57cb0 Parents: 30f55f2 Author: anoopsamjohnAuthored: Sat Nov 11 12:37:18 2017 +0530 Committer: anoopsamjohn Committed: Sat Nov 11 12:41:09 2017 +0530 -- .../hadoop/hbase/util/ByteBufferAllocator.java | 3 +- .../hadoop/hbase/util/ByteBufferArray.java | 18 +++ .../src/main/resources/hbase-default.xml| 7 +-- .../hadoop/hbase/util/TestByteBufferArray.java | 38 +- .../hbase/io/hfile/bucket/BucketCache.java | 6 +-- .../io/hfile/bucket/ByteBufferIOEngine.java | 18 ++- .../hbase/io/hfile/bucket/FileMmapEngine.java | 14 ++ .../hadoop/hbase/io/util/MemorySizeUtil.java| 44 +++- .../hbase/regionserver/HeapMemoryManager.java | 8 ++- ...estAvoidCellReferencesIntoShippedBlocks.java | 2 +- .../client/TestBlockEvictionFromClient.java | 2 +- .../hadoop/hbase/io/hfile/TestCacheConfig.java | 6 --- .../io/hfile/TestScannerFromBucketCache.java| 53 +++- .../hbase/io/hfile/bucket/TestBucketCache.java | 4 +- .../io/hfile/bucket/TestBucketWriterThread.java | 3 +- .../io/hfile/bucket/TestByteBufferIOEngine.java | 4 +- .../regionserver/TestHeapMemoryManager.java | 52 --- 17 files changed, 57 insertions(+), 225 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/3ee76f85/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java index 8e2ac16..2c5eac8 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java @@ -31,9 +31,8 @@ public interface ByteBufferAllocator { /** * Allocates a bytebuffer * @param size the size of the bytebuffer - * @param directByteBuffer indicator to create a direct bytebuffer * @return the bytebuffer that is created * @throws IOException exception thrown if there is an error while creating the ByteBuffer */ - ByteBuffer allocate(long size, boolean directByteBuffer) throws IOException; + ByteBuffer allocate(long size) throws IOException; } http://git-wip-us.apache.org/repos/asf/hbase/blob/3ee76f85/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java index d1d5e7d..2715740 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java @@ -59,11 +59,10 @@ public class ByteBufferArray { * of the array bounds for the last byte(see {@link ByteBufferArray#multiple}), * we will allocate one additional buffer with capacity 0; * @param capacity total size of the byte buffer array - * @param directByteBuffer true if we allocate direct buffer * @param allocator the ByteBufferAllocator that will create the buffers * @throws IOException throws IOException if there is an exception thrown by the allocator */ - public ByteBufferArray(long capacity, boolean directByteBuffer, ByteBufferAllocator allocator) + public ByteBufferArray(long capacity, ByteBufferAllocator allocator) throws IOException { this.bufferSize = DEFAULT_BUFFER_SIZE; if (this.bufferSize > (capacity / 16)) @@ -71,13 +70,13 @@ public class ByteBufferArray { this.bufferCount = (int) (roundUp(capacity, bufferSize) / bufferSize); LOG.info("Allocating buffers total=" + StringUtils.byteDesc(capacity) + ", sizePerBuffer=" + StringUtils.byteDesc(bufferSize) + ", count=" -+ bufferCount + ", direct=" + directByteBuffer); ++ bufferCount); buffers = new ByteBuffer[bufferCount + 1]; -createBuffers(directByteBuffer, allocator); +createBuffers(allocator); } @VisibleForTesting - void createBuffers(boolean directByteBuffer, ByteBufferAllocator allocator) + void
hbase git commit: HBASE-19187 Remove option to create on heap bucket cache.
Repository: hbase Updated Branches: refs/heads/master 72270866c -> bff619ef7 HBASE-19187 Remove option to create on heap bucket cache. Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/bff619ef Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/bff619ef Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/bff619ef Branch: refs/heads/master Commit: bff619ef7b100e8b09f7f5eb0f6e289ca51de096 Parents: 7227086 Author: anoopsamjohnAuthored: Sat Nov 11 12:37:18 2017 +0530 Committer: anoopsamjohn Committed: Sat Nov 11 12:37:18 2017 +0530 -- .../hadoop/hbase/util/ByteBufferAllocator.java | 3 +- .../hadoop/hbase/util/ByteBufferArray.java | 18 +++ .../src/main/resources/hbase-default.xml| 7 +-- .../hadoop/hbase/util/TestByteBufferArray.java | 38 +- .../hbase/io/hfile/bucket/BucketCache.java | 6 +-- .../io/hfile/bucket/ByteBufferIOEngine.java | 18 ++- .../hbase/io/hfile/bucket/FileMmapEngine.java | 14 ++ .../hadoop/hbase/io/util/MemorySizeUtil.java| 44 +++- .../hbase/regionserver/HeapMemoryManager.java | 8 ++- ...estAvoidCellReferencesIntoShippedBlocks.java | 2 +- .../client/TestBlockEvictionFromClient.java | 2 +- .../hadoop/hbase/io/hfile/TestCacheConfig.java | 6 --- .../io/hfile/TestScannerFromBucketCache.java| 53 +++- .../hbase/io/hfile/bucket/TestBucketCache.java | 4 +- .../io/hfile/bucket/TestBucketWriterThread.java | 3 +- .../io/hfile/bucket/TestByteBufferIOEngine.java | 4 +- .../regionserver/TestHeapMemoryManager.java | 52 --- 17 files changed, 57 insertions(+), 225 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/bff619ef/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java index 8e2ac16..2c5eac8 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferAllocator.java @@ -31,9 +31,8 @@ public interface ByteBufferAllocator { /** * Allocates a bytebuffer * @param size the size of the bytebuffer - * @param directByteBuffer indicator to create a direct bytebuffer * @return the bytebuffer that is created * @throws IOException exception thrown if there is an error while creating the ByteBuffer */ - ByteBuffer allocate(long size, boolean directByteBuffer) throws IOException; + ByteBuffer allocate(long size) throws IOException; } http://git-wip-us.apache.org/repos/asf/hbase/blob/bff619ef/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java -- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java index d1d5e7d..2715740 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteBufferArray.java @@ -59,11 +59,10 @@ public class ByteBufferArray { * of the array bounds for the last byte(see {@link ByteBufferArray#multiple}), * we will allocate one additional buffer with capacity 0; * @param capacity total size of the byte buffer array - * @param directByteBuffer true if we allocate direct buffer * @param allocator the ByteBufferAllocator that will create the buffers * @throws IOException throws IOException if there is an exception thrown by the allocator */ - public ByteBufferArray(long capacity, boolean directByteBuffer, ByteBufferAllocator allocator) + public ByteBufferArray(long capacity, ByteBufferAllocator allocator) throws IOException { this.bufferSize = DEFAULT_BUFFER_SIZE; if (this.bufferSize > (capacity / 16)) @@ -71,13 +70,13 @@ public class ByteBufferArray { this.bufferCount = (int) (roundUp(capacity, bufferSize) / bufferSize); LOG.info("Allocating buffers total=" + StringUtils.byteDesc(capacity) + ", sizePerBuffer=" + StringUtils.byteDesc(bufferSize) + ", count=" -+ bufferCount + ", direct=" + directByteBuffer); ++ bufferCount); buffers = new ByteBuffer[bufferCount + 1]; -createBuffers(directByteBuffer, allocator); +createBuffers(allocator); } @VisibleForTesting - void createBuffers(boolean directByteBuffer, ByteBufferAllocator allocator) + void
[2/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths
http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java index 2a0764d..5954e94 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java @@ -22,22 +22,24 @@ import static org.apache.hadoop.hbase.HConstants.DEFAULT_ZOOKEEPER_ZNODE_PARENT; import static org.apache.hadoop.hbase.HConstants.META_REPLICAS_NUM; import static org.apache.hadoop.hbase.HConstants.SPLIT_LOGDIR_NAME; import static org.apache.hadoop.hbase.HConstants.ZOOKEEPER_ZNODE_PARENT; -import static org.apache.hadoop.hbase.HRegionInfo.DEFAULT_REPLICA_ID; - -import org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap; +import static org.apache.hadoop.hbase.client.RegionInfo.DEFAULT_REPLICA_ID; import java.util.Optional; import java.util.stream.IntStream; import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.hbase.HRegionInfo; +import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap; + /** * Class that hold all the paths of znode for HBase. */ @InterfaceAudience.Private public class ZNodePaths { + // TODO: Replace this with ZooKeeper constant when ZOOKEEPER-277 is resolved. + public static final char ZNODE_PATH_SEPARATOR = '/'; public final static String META_ZNODE_PREFIX = "meta-region-server"; @@ -90,43 +92,35 @@ public class ZNodePaths { baseZNode = conf.get(ZOOKEEPER_ZNODE_PARENT, DEFAULT_ZOOKEEPER_ZNODE_PARENT); ImmutableMap.Builderbuilder = ImmutableMap.builder(); metaZNodePrefix = conf.get("zookeeper.znode.metaserver", META_ZNODE_PREFIX); -String defaultMetaReplicaZNode = ZKUtil.joinZNode(baseZNode, metaZNodePrefix); +String defaultMetaReplicaZNode = ZNodePaths.joinZNode(baseZNode, metaZNodePrefix); builder.put(DEFAULT_REPLICA_ID, defaultMetaReplicaZNode); int numMetaReplicas = conf.getInt(META_REPLICAS_NUM, DEFAULT_META_REPLICA_NUM); IntStream.range(1, numMetaReplicas) .forEachOrdered(i -> builder.put(i, defaultMetaReplicaZNode + "-" + i)); metaReplicaZNodes = builder.build(); -rsZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.rs", "rs")); -drainingZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.draining.rs", "draining")); -masterAddressZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.master", "master")); -backupMasterAddressesZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.backup.masters", "backup-masters")); -clusterStateZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.state", "running")); -tableZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.tableEnableDisable", "table")); -clusterIdZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.clusterId", "hbaseid")); -splitLogZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.splitlog", SPLIT_LOGDIR_NAME)); -balancerZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.balancer", "balancer")); -regionNormalizerZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.regionNormalizer", "normalizer")); -switchZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.switch", "switch")); -tableLockZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.tableLock", "table-lock")); -namespaceZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.namespace", "namespace")); -masterMaintZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.masterMaintenance", "master-maintenance")); -replicationZNode = -ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.replication", "replication")); +rsZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.rs", "rs")); +drainingZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.draining.rs", "draining")); +masterAddressZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.master", "master")); +backupMasterAddressesZNode = +joinZNode(baseZNode, conf.get("zookeeper.znode.backup.masters", "backup-masters")); +clusterStateZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.state", "running")); +tableZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.tableEnableDisable", "table")); +clusterIdZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.clusterId", "hbaseid")); +splitLogZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.splitlog",
[3/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths
HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths - Removes zookeeper connection from ClusterConnection - Deletes class ZooKeeperKeepAliveConnection - Removes Registry, ZooKeeperRegistry, and RegistryFactory Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/30f55f23 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/30f55f23 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/30f55f23 Branch: refs/heads/branch-2 Commit: 30f55f23166f3b0652bb2a96f8bbb1b2bcde4ace Parents: dd19fa2 Author: zhangduoAuthored: Fri Nov 10 22:16:45 2017 +0800 Committer: Apekshit Sharma Committed: Fri Nov 10 10:09:04 2017 -0800 -- .../hadoop/hbase/client/ClusterConnection.java | 4 + .../hbase/client/ConnectionImplementation.java | 236 +- .../apache/hadoop/hbase/client/HBaseAdmin.java | 326 +++ .../apache/hadoop/hbase/client/MetaCache.java | 27 +- .../apache/hadoop/hbase/client/Registry.java| 53 --- .../hadoop/hbase/client/RegistryFactory.java| 50 --- .../hadoop/hbase/client/ZKAsyncRegistry.java| 13 +- .../client/ZooKeeperKeepAliveConnection.java| 56 .../hadoop/hbase/client/ZooKeeperRegistry.java | 129 .../hbase/zookeeper/MasterAddressTracker.java | 2 +- .../hbase/zookeeper/RecoverableZooKeeper.java | 70 +--- .../hadoop/hbase/zookeeper/ZKMetadata.java | 80 + .../apache/hadoop/hbase/zookeeper/ZKUtil.java | 35 +- .../hadoop/hbase/zookeeper/ZNodePaths.java | 80 +++-- .../hadoop/hbase/zookeeper/ZkAclReset.java | 2 +- .../hbase/zookeeper/ZooKeeperWatcher.java | 2 +- .../hbase/client/DoNothingAsyncRegistry.java| 64 .../hadoop/hbase/client/TestAsyncProcess.java | 21 +- .../hbase/client/TestBufferedMutator.java | 33 +- .../hbase/client/TestClientNoCluster.java | 66 ++-- .../hbase/zookeeper/TestZooKeeperWatcher.java | 9 +- .../hbase/IntegrationTestMetaReplicas.java | 3 +- .../test/IntegrationTestZKAndFSPermissions.java | 3 +- .../replication/ReplicationPeersZKImpl.java | 3 +- .../ReplicationQueuesClientZKImpl.java | 9 +- .../replication/ReplicationQueuesZKImpl.java| 54 +-- .../replication/ReplicationStateZKBase.java | 17 +- .../hbase/rsgroup/RSGroupInfoManagerImpl.java | 11 +- .../rsgroup/VerifyingRSGroupAdminClient.java| 5 +- .../apache/hadoop/hbase/ZKNamespaceManager.java | 5 +- .../org/apache/hadoop/hbase/ZNodeClearer.java | 12 +- .../backup/example/HFileArchiveManager.java | 3 +- .../backup/example/ZKTableArchiveClient.java| 4 +- .../ZKSplitLogManagerCoordination.java | 10 +- .../ZkSplitLogWorkerCoordination.java | 9 +- .../hbase/master/ActiveMasterManager.java | 3 +- .../org/apache/hadoop/hbase/master/HMaster.java | 9 +- .../cleaner/ReplicationZKNodeCleaner.java | 7 +- .../hbase/procedure/ZKProcedureCoordinator.java | 5 +- .../hbase/procedure/ZKProcedureMemberRpcs.java | 10 +- .../hadoop/hbase/procedure/ZKProcedureUtil.java | 19 +- .../hbase/regionserver/HRegionServer.java | 3 +- .../security/access/ZKPermissionWatcher.java| 15 +- .../token/AuthenticationTokenSecretManager.java | 4 +- .../hbase/security/token/ZKSecretWatcher.java | 7 +- .../visibility/ZKVisibilityLabelWatcher.java| 5 +- .../org/apache/hadoop/hbase/util/HBaseFsck.java | 3 +- .../hadoop/hbase/util/ZKDataMigrator.java | 3 +- .../hbase/zookeeper/RegionServerTracker.java| 2 +- .../hbase/zookeeper/SplitOrMergeTracker.java| 4 +- .../hadoop/hbase/zookeeper/ZKSplitLog.java | 4 +- .../apache/hadoop/hbase/GenericTestUtils.java | 1 + .../hbase/client/TestMetaWithReplicas.java | 9 +- .../hbase/master/TestAssignmentListener.java| 5 +- .../hbase/master/TestMetaShutdownHandler.java | 3 +- .../hbase/master/TestTableStateManager.java | 3 +- .../TestCompactionInDeadRegionServer.java | 4 +- .../hbase/regionserver/TestSplitLogWorker.java | 7 +- .../replication/TestMasterReplication.java | 7 +- .../TestReplicationStateHBaseImpl.java | 3 +- .../replication/TestReplicationStateZKImpl.java | 9 +- .../TestReplicationTrackerZKImpl.java | 18 +- .../TestReplicationSourceManager.java | 8 +- .../token/TestZKSecretWatcherRefreshKeys.java | 3 +- .../hadoop/hbase/zookeeper/TestZKMulti.java | 40 +-- .../zookeeper/TestZooKeeperNodeTracker.java | 2 +- 66 files changed, 687 insertions(+), 1044 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java
[1/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths
Repository: hbase Updated Branches: refs/heads/branch-2 dd19fa25d -> 30f55f231 http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java index 9b5a44a..5a6a414 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java @@ -24,6 +24,7 @@ import org.apache.commons.logging.LogFactory; import org.apache.yetus.audience.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.zookeeper.ZKUtil; +import org.apache.hadoop.hbase.zookeeper.ZNodePaths; import org.apache.hadoop.hbase.zookeeper.ZooKeeperListener; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; @@ -53,8 +54,8 @@ public class ZKVisibilityLabelWatcher extends ZooKeeperListener { String labelZnodeParent = conf.get(VISIBILITY_LABEL_ZK_PATH, DEFAULT_VISIBILITY_LABEL_NODE); String userAuthsZnodeParent = conf.get(VISIBILITY_USER_AUTHS_ZK_PATH, DEFAULT_VISIBILITY_USER_AUTHS_NODE); -this.labelZnode = ZKUtil.joinZNode(watcher.znodePaths.baseZNode, labelZnodeParent); -this.userAuthsZnode = ZKUtil.joinZNode(watcher.znodePaths.baseZNode, userAuthsZnodeParent); +this.labelZnode = ZNodePaths.joinZNode(watcher.znodePaths.baseZNode, labelZnodeParent); +this.userAuthsZnode = ZNodePaths.joinZNode(watcher.znodePaths.baseZNode, userAuthsZnodeParent); } public void start() throws KeeperException { http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java index 934a630..bf4478d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java @@ -130,6 +130,7 @@ import org.apache.hadoop.hbase.wal.WALFactory; import org.apache.hadoop.hbase.wal.WALSplitter; import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; import org.apache.hadoop.hbase.zookeeper.ZKUtil; +import org.apache.hadoop.hbase.zookeeper.ZNodePaths; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException; import org.apache.hadoop.ipc.RemoteException; @@ -703,7 +704,7 @@ public class HBaseFsck extends Configured implements Closeable { */ private boolean setMasterInMaintenanceMode() throws IOException { RetryCounter retryCounter = createZNodeRetryCounterFactory.create(); -hbckEphemeralNodePath = ZKUtil.joinZNode( +hbckEphemeralNodePath = ZNodePaths.joinZNode( zkw.znodePaths.masterMaintZNode, "hbck-" + Long.toString(EnvironmentEdgeManager.currentTime())); do { http://git-wip-us.apache.org/repos/asf/hbase/blob/30f55f23/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java index 43c3598..e07861e 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java @@ -31,6 +31,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos; import org.apache.hadoop.hbase.zookeeper.ZKUtil; +import org.apache.hadoop.hbase.zookeeper.ZNodePaths; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; @@ -94,7 +95,7 @@ public class ZKDataMigrator { private static ZooKeeperProtos.DeprecatedTableState.State getTableState( final ZooKeeperWatcher zkw, final TableName tableName) throws KeeperException, InterruptedException { -String znode = ZKUtil.joinZNode(zkw.znodePaths.tableZNode, tableName.getNameAsString()); +String znode = ZNodePaths.joinZNode(zkw.znodePaths.tableZNode, tableName.getNameAsString()); byte [] data = ZKUtil.getData(zkw, znode); if (data == null || data.length <= 0) return null; try {
[1/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths
Repository: hbase Updated Branches: refs/heads/master 31234eb86 -> 72270866c http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java index 9b5a44a..5a6a414 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java @@ -24,6 +24,7 @@ import org.apache.commons.logging.LogFactory; import org.apache.yetus.audience.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.zookeeper.ZKUtil; +import org.apache.hadoop.hbase.zookeeper.ZNodePaths; import org.apache.hadoop.hbase.zookeeper.ZooKeeperListener; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; @@ -53,8 +54,8 @@ public class ZKVisibilityLabelWatcher extends ZooKeeperListener { String labelZnodeParent = conf.get(VISIBILITY_LABEL_ZK_PATH, DEFAULT_VISIBILITY_LABEL_NODE); String userAuthsZnodeParent = conf.get(VISIBILITY_USER_AUTHS_ZK_PATH, DEFAULT_VISIBILITY_USER_AUTHS_NODE); -this.labelZnode = ZKUtil.joinZNode(watcher.znodePaths.baseZNode, labelZnodeParent); -this.userAuthsZnode = ZKUtil.joinZNode(watcher.znodePaths.baseZNode, userAuthsZnodeParent); +this.labelZnode = ZNodePaths.joinZNode(watcher.znodePaths.baseZNode, labelZnodeParent); +this.userAuthsZnode = ZNodePaths.joinZNode(watcher.znodePaths.baseZNode, userAuthsZnodeParent); } public void start() throws KeeperException { http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java index 934a630..bf4478d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java @@ -130,6 +130,7 @@ import org.apache.hadoop.hbase.wal.WALFactory; import org.apache.hadoop.hbase.wal.WALSplitter; import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; import org.apache.hadoop.hbase.zookeeper.ZKUtil; +import org.apache.hadoop.hbase.zookeeper.ZNodePaths; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException; import org.apache.hadoop.ipc.RemoteException; @@ -703,7 +704,7 @@ public class HBaseFsck extends Configured implements Closeable { */ private boolean setMasterInMaintenanceMode() throws IOException { RetryCounter retryCounter = createZNodeRetryCounterFactory.create(); -hbckEphemeralNodePath = ZKUtil.joinZNode( +hbckEphemeralNodePath = ZNodePaths.joinZNode( zkw.znodePaths.masterMaintZNode, "hbck-" + Long.toString(EnvironmentEdgeManager.currentTime())); do { http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java index 43c3598..e07861e 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ZKDataMigrator.java @@ -31,6 +31,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos; import org.apache.hadoop.hbase.zookeeper.ZKUtil; +import org.apache.hadoop.hbase.zookeeper.ZNodePaths; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; import org.apache.zookeeper.KeeperException; @@ -94,7 +95,7 @@ public class ZKDataMigrator { private static ZooKeeperProtos.DeprecatedTableState.State getTableState( final ZooKeeperWatcher zkw, final TableName tableName) throws KeeperException, InterruptedException { -String znode = ZKUtil.joinZNode(zkw.znodePaths.tableZNode, tableName.getNameAsString()); +String znode = ZNodePaths.joinZNode(zkw.znodePaths.tableZNode, tableName.getNameAsString()); byte [] data = ZKUtil.getData(zkw, znode); if (data == null || data.length <= 0) return null; try {
[3/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths
HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths - Removes zookeeper connection from ClusterConnection - Deletes class ZooKeeperKeepAliveConnection - Removes Registry, ZooKeeperRegistry, and RegistryFactory Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/72270866 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/72270866 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/72270866 Branch: refs/heads/master Commit: 72270866cbacbcfd10f2ad2cce33362577eeef4d Parents: 31234eb Author: zhangduoAuthored: Fri Nov 10 22:16:45 2017 +0800 Committer: Apekshit Sharma Committed: Fri Nov 10 10:05:25 2017 -0800 -- .../hadoop/hbase/client/ClusterConnection.java | 4 + .../hbase/client/ConnectionImplementation.java | 236 +- .../apache/hadoop/hbase/client/HBaseAdmin.java | 326 +++ .../apache/hadoop/hbase/client/MetaCache.java | 27 +- .../apache/hadoop/hbase/client/Registry.java| 53 --- .../hadoop/hbase/client/RegistryFactory.java| 50 --- .../hadoop/hbase/client/ZKAsyncRegistry.java| 13 +- .../client/ZooKeeperKeepAliveConnection.java| 56 .../hadoop/hbase/client/ZooKeeperRegistry.java | 129 .../hbase/zookeeper/MasterAddressTracker.java | 2 +- .../hbase/zookeeper/RecoverableZooKeeper.java | 70 +--- .../hadoop/hbase/zookeeper/ZKMetadata.java | 80 + .../apache/hadoop/hbase/zookeeper/ZKUtil.java | 35 +- .../hadoop/hbase/zookeeper/ZNodePaths.java | 80 +++-- .../hadoop/hbase/zookeeper/ZkAclReset.java | 2 +- .../hbase/zookeeper/ZooKeeperWatcher.java | 2 +- .../hbase/client/DoNothingAsyncRegistry.java| 64 .../hadoop/hbase/client/TestAsyncProcess.java | 21 +- .../hbase/client/TestBufferedMutator.java | 33 +- .../hbase/client/TestClientNoCluster.java | 66 ++-- .../hbase/zookeeper/TestZooKeeperWatcher.java | 9 +- .../hbase/IntegrationTestMetaReplicas.java | 3 +- .../test/IntegrationTestZKAndFSPermissions.java | 3 +- .../replication/ReplicationPeersZKImpl.java | 3 +- .../ReplicationQueuesClientZKImpl.java | 9 +- .../replication/ReplicationQueuesZKImpl.java| 54 +-- .../replication/ReplicationStateZKBase.java | 17 +- .../hbase/rsgroup/RSGroupInfoManagerImpl.java | 11 +- .../rsgroup/VerifyingRSGroupAdminClient.java| 5 +- .../apache/hadoop/hbase/ZKNamespaceManager.java | 5 +- .../org/apache/hadoop/hbase/ZNodeClearer.java | 12 +- .../backup/example/HFileArchiveManager.java | 3 +- .../backup/example/ZKTableArchiveClient.java| 4 +- .../ZKSplitLogManagerCoordination.java | 10 +- .../ZkSplitLogWorkerCoordination.java | 9 +- .../hbase/master/ActiveMasterManager.java | 3 +- .../org/apache/hadoop/hbase/master/HMaster.java | 9 +- .../cleaner/ReplicationZKNodeCleaner.java | 7 +- .../hbase/procedure/ZKProcedureCoordinator.java | 5 +- .../hbase/procedure/ZKProcedureMemberRpcs.java | 10 +- .../hadoop/hbase/procedure/ZKProcedureUtil.java | 19 +- .../hbase/regionserver/HRegionServer.java | 3 +- .../security/access/ZKPermissionWatcher.java| 15 +- .../token/AuthenticationTokenSecretManager.java | 4 +- .../hbase/security/token/ZKSecretWatcher.java | 7 +- .../visibility/ZKVisibilityLabelWatcher.java| 5 +- .../org/apache/hadoop/hbase/util/HBaseFsck.java | 3 +- .../hadoop/hbase/util/ZKDataMigrator.java | 3 +- .../hbase/zookeeper/RegionServerTracker.java| 2 +- .../hbase/zookeeper/SplitOrMergeTracker.java| 4 +- .../hadoop/hbase/zookeeper/ZKSplitLog.java | 4 +- .../apache/hadoop/hbase/GenericTestUtils.java | 1 + .../hbase/client/TestMetaWithReplicas.java | 9 +- .../hbase/master/TestAssignmentListener.java| 5 +- .../hbase/master/TestMetaShutdownHandler.java | 3 +- .../hbase/master/TestTableStateManager.java | 3 +- .../TestCompactionInDeadRegionServer.java | 4 +- .../hbase/regionserver/TestSplitLogWorker.java | 7 +- .../replication/TestMasterReplication.java | 7 +- .../TestReplicationStateHBaseImpl.java | 3 +- .../replication/TestReplicationStateZKImpl.java | 9 +- .../TestReplicationTrackerZKImpl.java | 18 +- .../TestReplicationSourceManager.java | 8 +- .../token/TestZKSecretWatcherRefreshKeys.java | 3 +- .../hadoop/hbase/zookeeper/TestZKMulti.java | 40 +-- .../zookeeper/TestZooKeeperNodeTracker.java | 2 +- 66 files changed, 687 insertions(+), 1044 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClusterConnection.java
[2/3] hbase git commit: HBASE-19200 Make hbase-client only depend on ZKAsyncRegistry and ZNodePaths
http://git-wip-us.apache.org/repos/asf/hbase/blob/72270866/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java index 2a0764d..5954e94 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZNodePaths.java @@ -22,22 +22,24 @@ import static org.apache.hadoop.hbase.HConstants.DEFAULT_ZOOKEEPER_ZNODE_PARENT; import static org.apache.hadoop.hbase.HConstants.META_REPLICAS_NUM; import static org.apache.hadoop.hbase.HConstants.SPLIT_LOGDIR_NAME; import static org.apache.hadoop.hbase.HConstants.ZOOKEEPER_ZNODE_PARENT; -import static org.apache.hadoop.hbase.HRegionInfo.DEFAULT_REPLICA_ID; - -import org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap; +import static org.apache.hadoop.hbase.client.RegionInfo.DEFAULT_REPLICA_ID; import java.util.Optional; import java.util.stream.IntStream; import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.hbase.HRegionInfo; +import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap; + /** * Class that hold all the paths of znode for HBase. */ @InterfaceAudience.Private public class ZNodePaths { + // TODO: Replace this with ZooKeeper constant when ZOOKEEPER-277 is resolved. + public static final char ZNODE_PATH_SEPARATOR = '/'; public final static String META_ZNODE_PREFIX = "meta-region-server"; @@ -90,43 +92,35 @@ public class ZNodePaths { baseZNode = conf.get(ZOOKEEPER_ZNODE_PARENT, DEFAULT_ZOOKEEPER_ZNODE_PARENT); ImmutableMap.Builderbuilder = ImmutableMap.builder(); metaZNodePrefix = conf.get("zookeeper.znode.metaserver", META_ZNODE_PREFIX); -String defaultMetaReplicaZNode = ZKUtil.joinZNode(baseZNode, metaZNodePrefix); +String defaultMetaReplicaZNode = ZNodePaths.joinZNode(baseZNode, metaZNodePrefix); builder.put(DEFAULT_REPLICA_ID, defaultMetaReplicaZNode); int numMetaReplicas = conf.getInt(META_REPLICAS_NUM, DEFAULT_META_REPLICA_NUM); IntStream.range(1, numMetaReplicas) .forEachOrdered(i -> builder.put(i, defaultMetaReplicaZNode + "-" + i)); metaReplicaZNodes = builder.build(); -rsZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.rs", "rs")); -drainingZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.draining.rs", "draining")); -masterAddressZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.master", "master")); -backupMasterAddressesZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.backup.masters", "backup-masters")); -clusterStateZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.state", "running")); -tableZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.tableEnableDisable", "table")); -clusterIdZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.clusterId", "hbaseid")); -splitLogZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.splitlog", SPLIT_LOGDIR_NAME)); -balancerZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.balancer", "balancer")); -regionNormalizerZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.regionNormalizer", "normalizer")); -switchZNode = ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.switch", "switch")); -tableLockZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.tableLock", "table-lock")); -namespaceZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.namespace", "namespace")); -masterMaintZNode = ZKUtil.joinZNode(baseZNode, - conf.get("zookeeper.znode.masterMaintenance", "master-maintenance")); -replicationZNode = -ZKUtil.joinZNode(baseZNode, conf.get("zookeeper.znode.replication", "replication")); +rsZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.rs", "rs")); +drainingZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.draining.rs", "draining")); +masterAddressZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.master", "master")); +backupMasterAddressesZNode = +joinZNode(baseZNode, conf.get("zookeeper.znode.backup.masters", "backup-masters")); +clusterStateZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.state", "running")); +tableZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.tableEnableDisable", "table")); +clusterIdZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.clusterId", "hbaseid")); +splitLogZNode = joinZNode(baseZNode, conf.get("zookeeper.znode.splitlog",
hbase git commit: HBASE-16459 Remove unused hbase shell --format option
Repository: hbase Updated Branches: refs/heads/branch-1.1 f1e97d38b -> c4e3ad078 HBASE-16459 Remove unused hbase shell --format option Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c4e3ad07 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c4e3ad07 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c4e3ad07 Branch: refs/heads/branch-1.1 Commit: c4e3ad07898b42bf97d560ca1d2213276a4b9f9e Parents: f1e97d3 Author: Apekshit SharmaAuthored: Thu Nov 9 15:03:13 2017 -0800 Committer: Apekshit Sharma Committed: Fri Nov 10 09:55:11 2017 -0800 -- bin/hirb.rb | 17 + 1 file changed, 1 insertion(+), 16 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/c4e3ad07/bin/hirb.rb -- diff --git a/bin/hirb.rb b/bin/hirb.rb index 0503c29..03e9af3 100644 --- a/bin/hirb.rb +++ b/bin/hirb.rb @@ -48,30 +48,15 @@ $LOAD_PATH.unshift Pathname.new(sources) cmdline_help = <
hbase git commit: HBASE-16459 Remove unused hbase shell --format option
Repository: hbase Updated Branches: refs/heads/branch-1.2 ce5264a38 -> c2d154ca6 HBASE-16459 Remove unused hbase shell --format option Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/c2d154ca Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/c2d154ca Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/c2d154ca Branch: refs/heads/branch-1.2 Commit: c2d154ca652289608d91098da2a8fd541c8480c2 Parents: ce5264a Author: Apekshit SharmaAuthored: Thu Nov 9 15:03:13 2017 -0800 Committer: Apekshit Sharma Committed: Fri Nov 10 09:54:54 2017 -0800 -- bin/hirb.rb | 17 + 1 file changed, 1 insertion(+), 16 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/c2d154ca/bin/hirb.rb -- diff --git a/bin/hirb.rb b/bin/hirb.rb index 94b5cdb..470b093 100644 --- a/bin/hirb.rb +++ b/bin/hirb.rb @@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources) cmdline_help = <
hbase git commit: HBASE-16459 Remove unused hbase shell --format option
Repository: hbase Updated Branches: refs/heads/branch-1.3 96cad7314 -> d4acc9b68 HBASE-16459 Remove unused hbase shell --format option Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/d4acc9b6 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/d4acc9b6 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/d4acc9b6 Branch: refs/heads/branch-1.3 Commit: d4acc9b686c901f4b67a9b14966559ac12ab9f79 Parents: 96cad73 Author: Apekshit SharmaAuthored: Thu Nov 9 15:03:13 2017 -0800 Committer: Apekshit Sharma Committed: Fri Nov 10 09:54:40 2017 -0800 -- bin/hirb.rb | 17 + 1 file changed, 1 insertion(+), 16 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/d4acc9b6/bin/hirb.rb -- diff --git a/bin/hirb.rb b/bin/hirb.rb index 94b5cdb..470b093 100644 --- a/bin/hirb.rb +++ b/bin/hirb.rb @@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources) cmdline_help = <
hbase git commit: HBASE-16459 Remove unused hbase shell --format option
Repository: hbase Updated Branches: refs/heads/branch-1.4 6650a3ef8 -> 999cd0c3e HBASE-16459 Remove unused hbase shell --format option Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/999cd0c3 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/999cd0c3 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/999cd0c3 Branch: refs/heads/branch-1.4 Commit: 999cd0c3e8633312115fd568e2cdef0ff83e3f96 Parents: 6650a3e Author: Apekshit SharmaAuthored: Thu Nov 9 15:03:13 2017 -0800 Committer: Apekshit Sharma Committed: Fri Nov 10 09:54:00 2017 -0800 -- bin/hirb.rb | 17 + 1 file changed, 1 insertion(+), 16 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/999cd0c3/bin/hirb.rb -- diff --git a/bin/hirb.rb b/bin/hirb.rb index 2726e48..52343ca 100644 --- a/bin/hirb.rb +++ b/bin/hirb.rb @@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources) cmdline_help = <
hbase git commit: HBASE-16459 Remove unused hbase shell --format option
Repository: hbase Updated Branches: refs/heads/branch-1 696152657 -> 090ded568 HBASE-16459 Remove unused hbase shell --format option Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/090ded56 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/090ded56 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/090ded56 Branch: refs/heads/branch-1 Commit: 090ded5686a07b72f5d986c62d3b38e9c331df73 Parents: 6961526 Author: Apekshit SharmaAuthored: Thu Nov 9 15:03:13 2017 -0800 Committer: Apekshit Sharma Committed: Fri Nov 10 09:52:26 2017 -0800 -- bin/hirb.rb | 17 + 1 file changed, 1 insertion(+), 16 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/090ded56/bin/hirb.rb -- diff --git a/bin/hirb.rb b/bin/hirb.rb index 2726e48..52343ca 100644 --- a/bin/hirb.rb +++ b/bin/hirb.rb @@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources) cmdline_help = <
hbase git commit: HBASE-16459 Remove unused hbase shell --format option
Repository: hbase Updated Branches: refs/heads/branch-2 d757e651c -> dd19fa25d HBASE-16459 Remove unused hbase shell --format option Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/dd19fa25 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/dd19fa25 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/dd19fa25 Branch: refs/heads/branch-2 Commit: dd19fa25d506892f1f8f620c012b2a6ae56c1e03 Parents: d757e65 Author: Apekshit SharmaAuthored: Thu Nov 9 15:03:13 2017 -0800 Committer: Apekshit Sharma Committed: Fri Nov 10 00:34:16 2017 -0800 -- bin/hirb.rb | 17 + 1 file changed, 1 insertion(+), 16 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/dd19fa25/bin/hirb.rb -- diff --git a/bin/hirb.rb b/bin/hirb.rb index c06c83b..35895f8 100644 --- a/bin/hirb.rb +++ b/bin/hirb.rb @@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources) cmdline_help = <
[48/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/index-all.html -- diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html index 0a25ebf..317fb49 100644 --- a/devapidocs/index-all.html +++ b/devapidocs/index-all.html @@ -4084,6 +4084,8 @@ atLeast(TableSplit.Version) - Method in enum org.apache.hadoop.hbase.mapreduce.TableSplit.Version +atomic - Variable in class org.apache.hadoop.hbase.regionserver.HRegion.BatchOperation + ATOMIC_BOOLEAN - Static variable in class org.apache.hadoop.hbase.util.ClassSize Overhead for AtomicBoolean @@ -5535,6 +5537,8 @@ batchMutate(Mutation[], long, long) - Method in class org.apache.hadoop.hbase.regionserver.HRegion +batchMutate(Mutation[], boolean, long, long) - Method in class org.apache.hadoop.hbase.regionserver.HRegion + batchMutate(Mutation[]) - Method in class org.apache.hadoop.hbase.regionserver.HRegion batchMutate(HRegion.BatchOperation?) - Method in class org.apache.hadoop.hbase.regionserver.HRegion @@ -9104,6 +9108,8 @@ cellBuilder - Variable in class org.apache.hadoop.hbase.codec.MessageCodec.MessageDecoder +cellBuilder - Variable in class org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver + cellBuilder - Variable in class org.apache.hadoop.hbase.replication.BulkLoadCellFilter CellBuilder.DataType - Enum in org.apache.hadoop.hbase @@ -14103,6 +14109,8 @@ COMPARATOR - Static variable in interface org.apache.hadoop.hbase.client.TableDescriptor +comparator - Variable in class org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver + comparator - Variable in class org.apache.hadoop.hbase.filter.CompareFilter comparator - Variable in class org.apache.hadoop.hbase.filter.SingleColumnValueFilter @@ -24835,7 +24843,7 @@ doBatchMutate(Mutation) - Method in class org.apache.hadoop.hbase.regionserver.HRegion -doBatchOp(ClientProtos.RegionActionResult.Builder, HRegion, OperationQuota, ListClientProtos.Action, CellScanner, ActivePolicyEnforcement) - Method in class org.apache.hadoop.hbase.regionserver.RSRpcServices +doBatchOp(ClientProtos.RegionActionResult.Builder, HRegion, OperationQuota, ListClientProtos.Action, CellScanner, ActivePolicyEnforcement, boolean) - Method in class org.apache.hadoop.hbase.regionserver.RSRpcServices Execute a list of Put/Delete mutations. @@ -28926,6 +28934,10 @@ Size of the family length field in bytes +FAMILY_TO_ADD - Variable in class org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver + +FAMILY_TO_ADD_KEY - Static variable in class org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver + familyCellMaps - Variable in class org.apache.hadoop.hbase.regionserver.HRegion.BatchOperation familyClose(TRegion, TFamily) - Method in interface org.apache.hadoop.hbase.snapshot.SnapshotManifest.RegionVisitor @@ -29476,6 +29488,8 @@ FILES - Static variable in class org.apache.hadoop.hbase.snapshot.SnapshotInfo.Options +filesCompactedCounter - Variable in class org.apache.hadoop.hbase.coprocessor.example.ExampleRegionObserverWithMetrics + filesCompacting - Variable in class org.apache.hadoop.hbase.regionserver.HStore filesGroup - Variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot.ExportMapper @@ -30998,6 +31012,8 @@ flushConfigTable(MapString, RSGroupInfo) - Method in class org.apache.hadoop.hbase.rsgroup.RSGroupInfoManagerImpl +flushCounter - Variable in class org.apache.hadoop.hbase.coprocessor.example.ExampleRegionObserverWithMetrics + flushed - Variable in class org.apache.hadoop.hbase.codec.BaseEncoder flushed - Variable in class org.apache.hadoop.hbase.regionserver.StoreScanner @@ -44484,6 +44500,10 @@ getRegionObserver() - Method in class org.apache.hadoop.hbase.coprocessor.example.ExampleRegionObserverWithMetrics +getRegionObserver() - Method in class org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver + +getRegionObserver() - Method in class org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver + getRegionObserver() - Method in class org.apache.hadoop.hbase.coprocessor.example.WriteHeavyIncrementObserver getRegionObserver() - Method in class org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver @@ -58276,6 +58296,8 @@ isAsyncPrefetch() - Method in class org.apache.hadoop.hbase.client.Scan +isAtomic() - Method in class org.apache.hadoop.hbase.regionserver.HRegion.BatchOperation + isAuthorizationEnabled(Connection) - Static method in class org.apache.hadoop.hbase.security.access.AccessControlClient Return true if authorization is supported and enabled @@ -71914,7 +71936,7 @@ MUTATION_OVERHEAD - Static variable in class org.apache.hadoop.hbase.client.Mutation -MutationBatchOperation(HRegion, Mutation[], long, long) - Constructor for class
[47/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html index 32f538f..d9fb34d 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html @@ -114,60 +114,64 @@ -org.apache.hadoop.hbase.io +org.apache.hadoop.hbase.coprocessor.example +org.apache.hadoop.hbase.io + + + org.apache.hadoop.hbase.mapred Provides HBase http://wiki.apache.org/hadoop/HadoopMapReduce;>MapReduce Input/OutputFormats, a table indexing MapReduce job, and utility methods. - + org.apache.hadoop.hbase.mapreduce Provides HBase http://wiki.apache.org/hadoop/HadoopMapReduce;>MapReduce Input/OutputFormats, a table indexing MapReduce job, and utility methods. - + org.apache.hadoop.hbase.mapreduce.replication - + org.apache.hadoop.hbase.mob - + org.apache.hadoop.hbase.quotas - + org.apache.hadoop.hbase.regionserver - + org.apache.hadoop.hbase.regionserver.querymatcher - + org.apache.hadoop.hbase.rest.client - + org.apache.hadoop.hbase.rest.model - + org.apache.hadoop.hbase.security.access - + org.apache.hadoop.hbase.security.visibility - + org.apache.hadoop.hbase.thrift2 Provides an HBase http://thrift.apache.org/;>Thrift @@ -1202,6 +1206,25 @@ service. + + + +Uses of Scan in org.apache.hadoop.hbase.coprocessor.example + +Methods in org.apache.hadoop.hbase.coprocessor.example with parameters of type Scan + +Modifier and Type +Method and Description + + + +void +ScanModifyingObserver.preScannerOpen(ObserverContextRegionCoprocessorEnvironmentc, + Scanscan) + + + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/client/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html index 78ef84a..6955197 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html @@ -550,25 +550,25 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.client.RequestController.ReturnCode +org.apache.hadoop.hbase.client.HBaseAdmin.ReplicationState org.apache.hadoop.hbase.client.IsolationLevel +org.apache.hadoop.hbase.client.RequestController.ReturnCode org.apache.hadoop.hbase.client.MobCompactPartitionPolicy +org.apache.hadoop.hbase.client.CompactionState +org.apache.hadoop.hbase.client.Durability +org.apache.hadoop.hbase.client.ScannerCallable.MoreResults +org.apache.hadoop.hbase.client.AbstractResponse.ResponseType +org.apache.hadoop.hbase.client.AsyncRequestFutureImpl.Retry org.apache.hadoop.hbase.client.RegionLocateType -org.apache.hadoop.hbase.client.Consistency -org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState +org.apache.hadoop.hbase.client.CompactType org.apache.hadoop.hbase.client.AsyncProcessTask.SubmittedRows +org.apache.hadoop.hbase.client.Consistency org.apache.hadoop.hbase.client.MasterSwitchType -org.apache.hadoop.hbase.client.AbstractResponse.ResponseType -org.apache.hadoop.hbase.client.ScannerCallable.MoreResults -org.apache.hadoop.hbase.client.SnapshotType +org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState +org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState org.apache.hadoop.hbase.client.TableState.State -org.apache.hadoop.hbase.client.CompactType +org.apache.hadoop.hbase.client.SnapshotType org.apache.hadoop.hbase.client.Scan.ReadType -org.apache.hadoop.hbase.client.AsyncRequestFutureImpl.Retry -org.apache.hadoop.hbase.client.Durability -org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState -org.apache.hadoop.hbase.client.CompactionState -org.apache.hadoop.hbase.client.HBaseAdmin.ReplicationState http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/client/package-use.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-use.html b/devapidocs/org/apache/hadoop/hbase/client/package-use.html index
[19/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html index 8e0a6a2..08916af 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerState.html @@ -75,7 +75,7 @@ 067 }; 068 069 protected static final State[] STATES_EXPECTED_ON_CLOSE = new State[] { -070State.SPLITTING, State.SPLIT, // ServerCrashProcedure +070State.SPLITTING, State.SPLIT, State.MERGING, // ServerCrashProcedure 071State.OPEN, // enabled/open 072State.CLOSING // already in-progress (retrying) 073 }; @@ -550,446 +550,454 @@ 542return !getTableRegionStates(tableName).isEmpty(); 543 } 544 -545 public ListRegionInfo getRegionsOfTable(final TableName table) { -546return getRegionsOfTable(table, false); -547 } -548 -549 ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { -550final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); -551final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size()); -552for (RegionStateNode node: nodes) { -553 if (include(node, offline)) hris.add(node.getRegionInfo()); -554} -555return hris; -556 } -557 -558 /** -559 * Utility. Whether to include region in list of regions. Default is to -560 * weed out split and offline regions. -561 * @return True if we should include the codenode/code (do not include -562 * if split or offline unless codeoffline/code is set to true. -563 */ -564 boolean include(final RegionStateNode node, final boolean offline) { -565if (LOG.isTraceEnabled()) { -566 LOG.trace("WORKING ON " + node + " " + node.getRegionInfo()); -567} -568if (node.isInState(State.SPLIT)) return false; -569if (node.isInState(State.OFFLINE) !offline) return false; -570final RegionInfo hri = node.getRegionInfo(); -571return (!hri.isOffline() !hri.isSplit()) || -572((hri.isOffline() || hri.isSplit()) offline); -573 } -574 -575 /** -576 * Returns the set of regions hosted by the specified server -577 * @param serverName the server we are interested in -578 * @return set of RegionInfo hosted by the specified server -579 */ -580 public ListRegionInfo getServerRegionInfoSet(final ServerName serverName) { -581final ServerStateNode serverInfo = getServerNode(serverName); -582if (serverInfo == null) return Collections.emptyList(); -583 -584synchronized (serverInfo) { -585 return serverInfo.getRegionInfoList(); -586} -587 } -588 -589 // -590 // TODO: split helpers -591 // -592 public void logSplit(final ServerName serverName) { -593final ServerStateNode serverNode = getOrCreateServer(serverName); -594synchronized (serverNode) { -595 serverNode.setState(ServerState.SPLITTING); -596 /* THIS HAS TO BE WRONG. THIS IS SPLITTING OF REGION, NOT SPLITTING WALs. -597 for (RegionStateNode regionNode: serverNode.getRegions()) { -598synchronized (regionNode) { -599 // TODO: Abort procedure if present -600 regionNode.setState(State.SPLITTING); -601} -602 }*/ -603} -604 } -605 -606 public void logSplit(final RegionInfo regionInfo) { -607final RegionStateNode regionNode = getRegionNode(regionInfo); -608synchronized (regionNode) { -609 regionNode.setState(State.SPLIT); -610} -611 } -612 -613 @VisibleForTesting -614 public void updateRegionState(final RegionInfo regionInfo, final State state) { -615final RegionStateNode regionNode = getOrCreateRegionNode(regionInfo); +545 /** +546 * @return Return online regions of table; does not include OFFLINE or SPLITTING regions. +547 */ +548 public ListRegionInfo getRegionsOfTable(final TableName table) { +549return getRegionsOfTable(table, false); +550 } +551 +552 /** +553 * @return Return the regions of the table; does not include OFFLINE unless you set +554 * codeoffline/code to true. Does not include regions that are in the +555 * {@link State#SPLIT} state. +556 */ +557 public ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { +558final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); +559final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size());
[45/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html new file mode 100644 index 000..429b431 --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html @@ -0,0 +1,477 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +ScanModifyingObserver (Apache HBase 3.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.coprocessor.example +Class ScanModifyingObserver + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver + + + + + + + +All Implemented Interfaces: +Coprocessor, RegionCoprocessor, RegionObserver + + + +public class ScanModifyingObserver +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +implements RegionCoprocessor, RegionObserver +A RegionObserver which modifies incoming Scan requests to include additional + columns than what the user actually requested. + + + + + + + + + + + +Nested Class Summary + + + + +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.Coprocessor +Coprocessor.State + + + + + +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.coprocessor.RegionObserver +RegionObserver.MutationType + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +private byte[] +FAMILY_TO_ADD + + +static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +FAMILY_TO_ADD_KEY + + +private byte[] +QUALIFIER_TO_ADD + + +static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +QUALIFIER_TO_ADD_KEY + + + + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.Coprocessor +PRIORITY_HIGHEST, PRIORITY_LOWEST, PRIORITY_SYSTEM, PRIORITY_USER, VERSION + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +ScanModifyingObserver() + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalRegionObserver +getRegionObserver() + + +void +preScannerOpen(ObserverContextRegionCoprocessorEnvironmentc, + Scanscan) +Called before the client opens a new scanner. + + + +void +start(CoprocessorEnvironmentenv) +Called by the CoprocessorEnvironment during it's own startup to initialize the + coprocessor. + + + + + + + +Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode,
[36/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html index b77dc11..d1b724b 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":9,"i42":9,"i43":9,"i44":9,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":9,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":9,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":9,"i105":10,"i106":10,"i107":10,"i108":10,"i109":10 ,"i110":10,"i111":41,"i112":41,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":9,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":42,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":9,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":10,"i175":10,"i176":9,"i177":10,"i178":10,"i179":9,"i180":9,"i181":9,"i182":9,"i183":9,"i184":9,"i185":9,"i186":9,"i187":9,"i188":10,"i189":10,"i190":10,"i191":10,"i192":10,"i193":10,"i194":10,"i195":10,"i196":10,"i197":9,"i198":10,"i199":10,"i200":10,"i201":10,"i202":10,"i203":10,"i204":10,"i205":10,"i206":10,"i207":10,"i208":10,"i209":10,"i210":10 ,"i211":10,"i212":10,"i213":10,"i214":10,"i215":10,"i216":10,"i217":10,"i218":10,"i219":10,"i220":10,"i221":10,"i222":10,"i223":10,"i224":10,"i225":10,"i226":10,"i227":10,"i228":9,"i229":9,"i230":10,"i231":10,"i232":10,"i233":10,"i234":10,"i235":10,"i236":10,"i237":10,"i238":10,"i239":10,"i240":10,"i241":9,"i242":10,"i243":10,"i244":10,"i245":10,"i246":10,"i247":10,"i248":10,"i249":10,"i250":10,"i251":10,"i252":10,"i253":10,"i254":10,"i255":10,"i256":9,"i257":10,"i258":10,"i259":10,"i260":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":9,"i43":9,"i44":9,"i45":9,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":9,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":9,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":9,"i106":10,"i107":10,"i108":10,"i109":10
[46/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html index 1d1a3ec..9fd977c 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -class ExampleRegionObserverWithMetrics.ExampleRegionObserver +class ExampleRegionObserverWithMetrics.ExampleRegionObserver extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements RegionCoprocessor, RegionObserver @@ -202,13 +202,40 @@ implements void +postCompactSelection(ObserverContextRegionCoprocessorEnvironmentc, +Storestore, +http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends StoreFileselected, +CompactionLifeCycleTrackertracker, +CompactionRequestrequest) +Called after the StoreFiles to compact have been selected from the available + candidates. + + + +void +postFlush(ObserverContextRegionCoprocessorEnvironmentc, + FlushLifeCycleTrackertracker) +Called after the memstore is flushed to disk. + + + +void +postFlush(ObserverContextRegionCoprocessorEnvironmentc, + Storestore, + StoreFileresultFile, + FlushLifeCycleTrackertracker) +Called after a Store's memstore is flushed to disk. + + + +void postGetOp(ObserverContextRegionCoprocessorEnvironmente, Getget, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCellresults) Called after the client performs a Get - + void preGetOp(ObserverContextRegionCoprocessorEnvironmente, Getget, @@ -243,7 +270,7 @@ implements RegionObserver -postAppend, postBatchMutate, postBatchMutateIndispensably, postBulkLoadHFile, postCheckAndDelete, postCheckAndPut, postClose, postCloseRegionOperation, postCommitStoreFile, postCompact, postCompactSelection, postDelete, postExists, postFlush, postFlush, postIncrement, postInstantiateDeleteTracker, postMemStoreCompaction, postMutationBeforeWAL, postOpen, postPut, postReplayWALs, postScannerClose, postScannerFilterRow, postScannerNext, postScannerOpen, postStartRegionOperation, postStoreFileReaderOpen, postWALRestore, preAppend, preAppendAfterRowLock, preBatchMutate, preBulkLoadHFile, preCheckAndDelete, preCheckAndDeleteAfterRowLock, preCheckAndPut, preCheckAndPutAfterRowLock, preClose, preCommitStoreFile, preCompact, preCompactScannerOpen, preCompactSelection, preDelete, preExists, preFlush, preFlush, preFlushScannerOpen, preIncrement, preIncrementAfterRowLock, preMemStoreCompaction, preMemStoreCompactionCompact, preMemStoreCompactionCompactScannerOpen, preOpen, prePrepareTimeStampForDeleteVersion, prePut, preReplayWALs, preScannerClose, preScannerNext, preScannerOpen, preStoreFileReaderOpen, preStoreScannerOpen, preWALRestore +postAppend, postBatchMutate, postBatchMutateIndispensably, postBulkLoadHFile, postCheckAndDelete, postCheckAndPut, postClose, postCloseRegionOperation, postCommitStoreFile, postCompact, postDelete, postExists, postIncrement, postInstantiateDeleteTracker, postMemStoreCompaction, postMutationBeforeWAL, postOpen, postPut, postReplayWALs, postScannerClose, postScannerFilterRow, postScannerNext, postScannerOpen, postStartRegionOperation, postStoreFileReaderOpen, postWALRestore, preAppend, preAppendAfterRowLock, preBatchMutate, preBulkLoadHFile, preCheckAndDelete, preCheckAndDeleteAfterRowLock, preCheckAndPut, preCheckAndPutAfterRowLock, preClose, preCommitStoreFile, pre Compact, preCompactScannerOpen, preCompactSelection, preDelete, preExists, preFlush, preFlush, preFlushScannerOpen, preIncrement, preIncrementAfterRowLock, preMemStoreCompaction, preMemStoreCompactionCompact, preMemStoreCompactionCompactScannerOpen, preOpen, prePrepareTimeStampForDeleteVersion, prePut,
[44/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html new file mode 100644 index 000..075273c --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html @@ -0,0 +1,515 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +ValueRewritingObserver (Apache HBase 3.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.coprocessor.example +Class ValueRewritingObserver + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver + + + + + + + +All Implemented Interfaces: +Coprocessor, RegionCoprocessor, RegionObserver + + + +public class ValueRewritingObserver +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +implements RegionObserver, RegionCoprocessor +This RegionObserver replaces the values of Puts from one value to another on compaction. + + + + + + + + + + + +Nested Class Summary + + + + +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.coprocessor.RegionObserver +RegionObserver.MutationType + + + + + +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.Coprocessor +Coprocessor.State + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +private CellBuilder +cellBuilder + + +private Bytes.ByteArrayComparator +comparator + + +static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +ORIGINAL_VALUE_KEY + + +static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +REPLACED_VALUE_KEY + + +private byte[] +replacedValue + + +private byte[] +sourceValue + + + + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.Coprocessor +PRIORITY_HIGHEST, PRIORITY_LOWEST, PRIORITY_SYSTEM, PRIORITY_USER, VERSION + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +ValueRewritingObserver() + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalRegionObserver +getRegionObserver() + + +InternalScanner +preCompact(ObserverContextRegionCoprocessorEnvironmentc, + Storestore, + InternalScannerscanner, + ScanTypescanType, + CompactionLifeCycleTrackertracker, + CompactionRequestrequest) +Called prior to writing the StoreFiles selected for compaction into a new + StoreFile. + + + +void +start(CoprocessorEnvironmentenv) +Called by the CoprocessorEnvironment during it's own startup to initialize the + coprocessor. + + + + + + + +Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize,
[37/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html index 5f09cd0..5833858 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -public static class HRegion.RowLockImpl +public static class HRegion.RowLockImpl extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements Region.RowLock Class used to represent a lock on a row. @@ -226,7 +226,7 @@ implements context -private finalHRegion.RowLockContext context +private finalHRegion.RowLockContext context @@ -235,7 +235,7 @@ implements lock -private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">Lock lock +private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">Lock lock @@ -252,7 +252,7 @@ implements RowLockImpl -publicRowLockImpl(HRegion.RowLockContextcontext, +publicRowLockImpl(HRegion.RowLockContextcontext, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">Locklock) @@ -270,7 +270,7 @@ implements getLock -publichttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">LockgetLock() +publichttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">LockgetLock() @@ -279,7 +279,7 @@ implements getContext -publicHRegion.RowLockContextgetContext() +publicHRegion.RowLockContextgetContext() @@ -288,7 +288,7 @@ implements release -publicvoidrelease() +publicvoidrelease() Description copied from interface:Region.RowLock Release the given lock. If there are no remaining locks held by the current thread then unlock the row and allow other threads to acquire the lock. @@ -304,7 +304,7 @@ implements toString -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() Overrides: http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--; title="class or interface in java.lang">toStringin classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object
[50/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/checkstyle-aggregate.html -- diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html index 3d0abd1..3fa0a8f 100644 --- a/checkstyle-aggregate.html +++ b/checkstyle-aggregate.html @@ -7,7 +7,7 @@ - + Apache HBase Checkstyle Results @@ -286,10 +286,10 @@ Warnings Errors -3422 +3426 0 0 -21453 +21437 Files @@ -6262,7 +6262,7 @@ org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.java 0 0 -84 +75 org/apache/hadoop/hbase/master/assignment/MockMasterServices.java 0 @@ -6277,7 +6277,7 @@ org/apache/hadoop/hbase/master/assignment/RegionStateStore.java 0 0 -10 +9 org/apache/hadoop/hbase/master/assignment/RegionStates.java 0 @@ -8377,7 +8377,7 @@ org/apache/hadoop/hbase/regionserver/RSRpcServices.java 0 0 -142 +136 org/apache/hadoop/hbase/regionserver/RSStatusServlet.java 0 @@ -13199,12 +13199,12 @@ http://checkstyle.sourceforge.net/config_blocks.html#NeedBraces;>NeedBraces -2281 +2280 Error coding http://checkstyle.sourceforge.net/config_coding.html#EmptyStatement;>EmptyStatement -56 +55 Error @@ -13284,7 +13284,7 @@ caseIndent: 2 basicOffset: 2 lineWrappingIndentation: 2 -6309 +6300 Error javadoc @@ -13296,7 +13296,7 @@ http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription;>NonEmptyAtclauseDescription -4355 +4351 Error misc @@ -13314,7 +13314,7 @@ max: 100 ignorePattern: ^package.*|^import.*|a href|href|http://|https://|ftp://|org.apache.thrift.|com.google.protobuf.|hbase.protobuf.generated -1952 +1951 Error @@ -84053,91 +84053,91 @@ blocks NeedBraces 'if' construct must use '{}'s. -1327 +1336 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -1345 +1354 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -1347 +1356 Error blocks NeedBraces 'if' construct must use '{}'s. -1350 +1359 Error blocks NeedBraces 'if' construct must use '{}'s. -1355 +1364 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -1363 +1372 Error blocks NeedBraces 'if' construct must use '{}'s. -1368 +1377 Error blocks NeedBraces 'if' construct must use '{}'s. -1373 +1382 Error blocks NeedBraces 'else' construct must use '{}'s. -1374 +1383 Error blocks EmptyBlock Must have at least one statement. -1455 +1464 Error sizes LineLength Line is longer than 100 characters (found 101). -1482 +1491 Error blocks NeedBraces 'if' construct must use '{}'s. -1638 +1652 Error blocks NeedBraces 'if' construct must use '{}'s. -1765 +1779 Error indentation Indentation 'lambda arguments' have incorrect indentation level 8, expected level should be 6. -1810 +1824 Error blocks NeedBraces 'if' construct must use '{}'s. -1827 +1841 org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.java @@ -84452,515 +84452,461 @@ imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting' import. -67 +68 Error indentation Indentation 'case' child have incorrect indentation level 6, expected level should be 8. -216 +217 Error indentation Indentation 'if' have incorrect indentation level 8, expected level should be 10. -217 +218 Error indentation Indentation 'if' child have incorrect indentation level 10, expected level should be 12. -219 +220 Error indentation Indentation 'if rcurly' have incorrect indentation level 8, expected level should be 10. -220 - -Error -indentation -Indentation -'block' child have incorrect indentation level 8, expected level should be 10. 221 - -Error -indentation -Indentation -'block' child have incorrect indentation level 8, expected level should be 10. -222 - -Error -indentation -Indentation -'case' child have incorrect indentation level 6, expected level should be 8. -223 - -Error -indentation -Indentation -'block' child have incorrect indentation level 8, expected level should be 10. -224 Error indentation Indentation 'block' child have incorrect indentation level 8, expected level should be 10. -225 +222 Error indentation Indentation 'block' child have incorrect indentation level 8, expected level should be 10. -226 +223 Error indentation Indentation 'case' child have incorrect indentation level 6, expected level should be 8. -227 +224 Error indentation Indentation 'block' child have incorrect indentation level 8, expected level should be 10. -228 +225 Error indentation Indentation 'block' child have incorrect indentation level 8, expected level should be 10. -229 +226 Error indentation Indentation 'block' child have incorrect indentation level 8, expected level should be 10. -230 +227 Error
[42/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html index a749d86..c80374f 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html @@ -129,7 +129,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class MergeTableRegionsProcedure +public class MergeTableRegionsProcedure extends AbstractStateMachineTableProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsState The procedure to Merge a region in a table. This procedure takes an exclusive table lock since it is working over multiple regions. @@ -483,15 +483,16 @@ extends -void -setRegionStateToMerging(MasterProcedureEnvenv) -Set the region states to MERGING state +private void +setRegionStateBackToOpen(MasterProcedureEnvenv) +Rollback the region state change + Not used for now, since rollbackCloseRegionsForMerge() will mark regions as OPEN -private void -setRegionStateToRevertMerging(MasterProcedureEnvenv) -Rollback the region state change +void +setRegionStateToMerging(MasterProcedureEnvenv) +Set the region states to MERGING state @@ -556,7 +557,7 @@ extends LOG -private static finalorg.apache.commons.logging.Log LOG +private static finalorg.apache.commons.logging.Log LOG @@ -565,7 +566,7 @@ extends traceEnabled -privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled +privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled @@ -574,7 +575,7 @@ extends lock -private volatileboolean lock +private volatileboolean lock @@ -583,7 +584,7 @@ extends regionLocation -privateServerName regionLocation +privateServerName regionLocation @@ -592,7 +593,7 @@ extends regionsToMerge -privateRegionInfo[] regionsToMerge +privateRegionInfo[] regionsToMerge @@ -601,7 +602,7 @@ extends mergedRegion -privateRegionInfo mergedRegion +privateRegionInfo mergedRegion @@ -610,7 +611,7 @@ extends forcible -privateboolean forcible +privateboolean forcible @@ -627,7 +628,7 @@ extends MergeTableRegionsProcedure -publicMergeTableRegionsProcedure() +publicMergeTableRegionsProcedure() @@ -636,7 +637,7 @@ extends MergeTableRegionsProcedure -publicMergeTableRegionsProcedure(MasterProcedureEnvenv, +publicMergeTableRegionsProcedure(MasterProcedureEnvenv, RegionInforegionToMergeA, RegionInforegionToMergeB) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -652,7 +653,7 @@ extends MergeTableRegionsProcedure -publicMergeTableRegionsProcedure(MasterProcedureEnvenv, +publicMergeTableRegionsProcedure(MasterProcedureEnvenv, RegionInforegionToMergeA, RegionInforegionToMergeB, booleanforcible) @@ -669,7 +670,7 @@ extends MergeTableRegionsProcedure -publicMergeTableRegionsProcedure(MasterProcedureEnvenv, +publicMergeTableRegionsProcedure(MasterProcedureEnvenv, RegionInfo[]regionsToMerge, booleanforcible) throws MergeRegionException @@ -693,7 +694,7 @@ extends checkRegionsToMerge -private staticvoidcheckRegionsToMerge(RegionInfo[]regionsToMerge, +private staticvoidcheckRegionsToMerge(RegionInfo[]regionsToMerge, booleanforcible) throws MergeRegionException @@ -708,7 +709,7 @@ extends checkRegionsToMerge -private staticvoidcheckRegionsToMerge(RegionInforegionToMergeA, +private staticvoidcheckRegionsToMerge(RegionInforegionToMergeA, RegionInforegionToMergeB, booleanforcible) throws MergeRegionException @@ -724,7 +725,7 @@ extends createMergedRegionInfo -private staticRegionInfocreateMergedRegionInfo(RegionInfo[]regionsToMerge) +private staticRegionInfocreateMergedRegionInfo(RegionInfo[]regionsToMerge) @@ -733,7 +734,7 @@ extends createMergedRegionInfo -private staticRegionInfocreateMergedRegionInfo(RegionInforegionToMergeA, +private
hbase-site git commit: INFRA-10751 Empty commit
Repository: hbase-site Updated Branches: refs/heads/asf-site a108018f1 -> bac2e663f INFRA-10751 Empty commit Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/bac2e663 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/bac2e663 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/bac2e663 Branch: refs/heads/asf-site Commit: bac2e663fcd1745673991e367bab24750e970ab6 Parents: a108018 Author: jenkinsAuthored: Fri Nov 10 15:30:07 2017 + Committer: jenkins Committed: Fri Nov 10 15:30:07 2017 + -- --
[43/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html index d4bc51c..efa1c09 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html @@ -406,7 +406,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Setbyte[]regionNames) -void +(package private) void checkOnlineRegionsReportForMeta(RegionStates.ServerStateNodeserverNode, http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Setbyte[]regionNames) @@ -1299,7 +1299,7 @@ implements pendingAssignQueue -private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true; title="class or interface in java.util">ArrayListRegionStates.RegionStateNode pendingAssignQueue +private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true; title="class or interface in java.util">ArrayListRegionStates.RegionStateNode pendingAssignQueue @@ -1308,7 +1308,7 @@ implements assignQueueLock -private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/ReentrantLock.html?is-external=true; title="class or interface in java.util.concurrent.locks">ReentrantLock assignQueueLock +private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/ReentrantLock.html?is-external=true; title="class or interface in java.util.concurrent.locks">ReentrantLock assignQueueLock @@ -1317,7 +1317,7 @@ implements assignQueueFullCond -private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Condition.html?is-external=true; title="class or interface in java.util.concurrent.locks">Condition assignQueueFullCond +private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Condition.html?is-external=true; title="class or interface in java.util.concurrent.locks">Condition assignQueueFullCond @@ -2126,8 +2126,8 @@ implements checkOnlineRegionsReportForMeta -publicvoidcheckOnlineRegionsReportForMeta(RegionStates.ServerStateNodeserverNode, -http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Setbyte[]regionNames) +voidcheckOnlineRegionsReportForMeta(RegionStates.ServerStateNodeserverNode, + http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Setbyte[]regionNames) @@ -2225,7 +2225,7 @@ implements processofflineServersWithOnlineRegions -privatebooleanprocessofflineServersWithOnlineRegions() +privatebooleanprocessofflineServersWithOnlineRegions() @@ -2234,7 +2234,7 @@ implements isFailoverCleanupDone -publicbooleanisFailoverCleanupDone() +publicbooleanisFailoverCleanupDone() Used by ServerCrashProcedure to make sure AssignmentManager has completed the failover cleanup before re-assigning regions of dead servers. So that when re-assignment happens, AssignmentManager has proper region states. @@ -2246,7 +2246,7 @@ implements setFailoverCleanupDone -publicvoidsetFailoverCleanupDone(booleanb) +publicvoidsetFailoverCleanupDone(booleanb) Used by ServerCrashProcedure tests verify the ability to suspend the execution of the ServerCrashProcedure. @@ -2257,7 +2257,7 @@ implements getFailoverCleanupEvent -publicProcedureEventgetFailoverCleanupEvent() +publicProcedureEventgetFailoverCleanupEvent() @@ -2266,7 +2266,7 @@ implements checkFailoverCleanupCompleted -privatevoidcheckFailoverCleanupCompleted(RegionInfohri) +privatevoidcheckFailoverCleanupCompleted(RegionInfohri) throws PleaseHoldException Used to check if the failover cleanup is done. if not we throw PleaseHoldException since we are rebuilding the RegionStates @@ -2284,7 +2284,7 @@ implements getNumRegionsOpened -publicintgetNumRegionsOpened() +publicintgetNumRegionsOpened() @@ -2293,7 +2293,7 @@ implements submitServerCrash -publicvoidsubmitServerCrash(ServerNameserverName, +publicvoidsubmitServerCrash(ServerNameserverName, booleanshouldSplitWal) @@ -2303,7 +2303,7 @@ implements offlineRegion -publicvoidofflineRegion(RegionInforegionInfo) +publicvoidofflineRegion(RegionInforegionInfo) @@ -2312,7 +2312,7 @@ implements onlineRegion -publicvoidonlineRegion(RegionInforegionInfo,
[49/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index ad53a23..e3b0bd0 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -25,8 +25,8 @@ under the License. en-us 2007 - 2017 The Apache Software Foundation - File: 3422, - Errors: 21453, + File: 3426, + Errors: 21437, Warnings: 0, Infos: 0 @@ -7704,6 +7704,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.example.TestScanModifyingObserver.java;>org/apache/hadoop/hbase/coprocessor/example/TestScanModifyingObserver.java + + + 0 + + + 0 + + + 0 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.AbstractByteRange.java;>org/apache/hadoop/hbase/util/AbstractByteRange.java @@ -14942,6 +14956,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.example.ValueRewritingObserver.java;>org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.java + + + 0 + + + 0 + + + 0 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.CommonFSUtils.java;>org/apache/hadoop/hbase/util/CommonFSUtils.java @@ -18339,7 +18367,7 @@ under the License. 0 - 142 + 136 @@ -20789,7 +20817,7 @@ under the License. 0 - 10 + 9 @@ -21349,7 +21377,7 @@ under the License. 0 - 84 + 75 @@ -35018,6 +35046,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.example.TestValueReplacingCompaction.java;>org/apache/hadoop/hbase/coprocessor/example/TestValueReplacingCompaction.java + + + 0 + + + 0 + + + 0 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.StripeStoreConfig.java;>org/apache/hadoop/hbase/regionserver/StripeStoreConfig.java @@ -40534,6 +40576,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.example.ScanModifyingObserver.java;>org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.java + + + 0 + + + 0 + + + 0 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.client.example.RefreshHFilesClient.java;>org/apache/hadoop/hbase/client/example/RefreshHFilesClient.java http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/coc.html -- diff --git a/coc.html b/coc.html index 51ce016..14a8353 100644 --- a/coc.html +++ b/coc.html @@ -7,7 +7,7 @@ - + Apache HBase Code of Conduct Policy @@ -380,7 +380,7 @@ email to mailto:priv...@hbase.apache.org;>the priv https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-11-09 + Last Published: 2017-11-10 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/cygwin.html
[38/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html index 24779ca..e04d11d 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html @@ -121,7 +121,7 @@ var activeTableTab = "activeTableTab"; -class HRegion.RegionScannerImpl +class HRegion.RegionScannerImpl extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements RegionScanner, Shipper, RpcCallback RegionScannerImpl is used to combine scanners from multiple Stores (aka column families). @@ -425,7 +425,7 @@ implements storeHeap -KeyValueHeap storeHeap +KeyValueHeap storeHeap @@ -434,7 +434,7 @@ implements joinedHeap -KeyValueHeap joinedHeap +KeyValueHeap joinedHeap Heap of key-values that are not essential for the provided filters and are thus read on demand, if on-demand column family loading is enabled. @@ -445,7 +445,7 @@ implements joinedContinuationRow -protectedCell joinedContinuationRow +protectedCell joinedContinuationRow If the joined heap data gathering is interrupted due to scan limits, this will contain the row for which we are populating the values. @@ -456,7 +456,7 @@ implements filterClosed -privateboolean filterClosed +privateboolean filterClosed @@ -465,7 +465,7 @@ implements stopRow -protected finalbyte[] stopRow +protected finalbyte[] stopRow @@ -474,7 +474,7 @@ implements includeStopRow -protected finalboolean includeStopRow +protected finalboolean includeStopRow @@ -483,7 +483,7 @@ implements region -protected finalHRegion region +protected finalHRegion region @@ -492,7 +492,7 @@ implements comparator -protected finalCellComparator comparator +protected finalCellComparator comparator @@ -501,7 +501,7 @@ implements readPt -private finallong readPt +private finallong readPt @@ -510,7 +510,7 @@ implements maxResultSize -private finallong maxResultSize +private finallong maxResultSize @@ -519,7 +519,7 @@ implements defaultScannerContext -private finalScannerContext defaultScannerContext +private finalScannerContext defaultScannerContext @@ -528,7 +528,7 @@ implements filter -private finalFilterWrapper filter +private finalFilterWrapper filter @@ -545,7 +545,7 @@ implements RegionScannerImpl -RegionScannerImpl(Scanscan, +RegionScannerImpl(Scanscan, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListKeyValueScanneradditionalScanners, HRegionregion) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -561,7 +561,7 @@ implements RegionScannerImpl -RegionScannerImpl(Scanscan, +RegionScannerImpl(Scanscan, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListKeyValueScanneradditionalScanners, HRegionregion, longnonceGroup, @@ -587,7 +587,7 @@ implements getRegionInfo -publicRegionInfogetRegionInfo() +publicRegionInfogetRegionInfo() Specified by: getRegionInfoin interfaceRegionScanner @@ -602,7 +602,7 @@ implements initializeScanners -protectedvoidinitializeScanners(Scanscan, +protectedvoidinitializeScanners(Scanscan, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListKeyValueScanneradditionalScanners) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -617,7 +617,7 @@ implements initializeKVHeap -protectedvoidinitializeKVHeap(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListKeyValueScannerscanners, +protectedvoidinitializeKVHeap(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListKeyValueScannerscanners, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListKeyValueScannerjoinedScanners, HRegionregion) throws
[32/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html index dc2f127..91776e4 100644 --- a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html @@ -110,8 +110,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.ByteArrayComparableModel.ComparatorType org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.FilterType +org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.ByteArrayComparableModel.ComparatorType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html b/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html index d2946f4..793c33c 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html @@ -136,9 +136,9 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.security.access.AccessController.OpType -org.apache.hadoop.hbase.security.access.Permission.Action org.apache.hadoop.hbase.security.access.AccessControlFilter.Strategy +org.apache.hadoop.hbase.security.access.Permission.Action +org.apache.hadoop.hbase.security.access.AccessController.OpType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/security/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html index c4d3f6e..bec1bbc 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html @@ -191,8 +191,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection org.apache.hadoop.hbase.security.AuthMethod +org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection org.apache.hadoop.hbase.security.SaslStatus http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html index 29f4bfc..59ab0ce 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html @@ -198,9 +198,9 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html index cb99dd4..e7374e1 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html @@ -55,739 +55,744 @@ 047import org.apache.hadoop.hbase.master.MasterCoprocessorHost; 048import org.apache.hadoop.hbase.master.MasterFileSystem; 049import org.apache.hadoop.hbase.master.RegionState; -050import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -051import org.apache.hadoop.hbase.master.procedure.AbstractStateMachineTableProcedure; -052import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; -053import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil; -054import org.apache.hadoop.hbase.procedure2.ProcedureMetrics; -055import org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer; -056import org.apache.hadoop.hbase.procedure2.ProcedureSuspendedException; -057import org.apache.hadoop.hbase.procedure2.ProcedureYieldException; -058import org.apache.hadoop.hbase.quotas.QuotaExceededException; -059import org.apache.hadoop.hbase.regionserver.HRegionFileSystem; -060import org.apache.hadoop.hbase.regionserver.HStoreFile; -061import org.apache.hadoop.hbase.regionserver.StoreFileInfo; -062import org.apache.hadoop.hbase.util.Bytes; -063import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -064import org.apache.hadoop.hbase.util.FSUtils; -065import org.apache.yetus.audience.InterfaceAudience; -066 -067import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting; -068import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -069import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -070import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos; -071import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsState; -072 -073/** -074 * The procedure to Merge a region in a table. -075 * This procedure takes an exclusive table lock since it is working over multiple regions. -076 * It holds the lock for the life of the procedure. -077 */ -078@InterfaceAudience.Private -079public class MergeTableRegionsProcedure -080extends AbstractStateMachineTableProcedureMergeTableRegionsState { -081 private static final Log LOG = LogFactory.getLog(MergeTableRegionsProcedure.class); -082 private Boolean traceEnabled; -083 private volatile boolean lock = false; -084 private ServerName regionLocation; -085 private RegionInfo[] regionsToMerge; -086 private RegionInfo mergedRegion; -087 private boolean forcible; -088 -089 public MergeTableRegionsProcedure() { -090// Required by the Procedure framework to create the procedure on replay -091 } -092 -093 public MergeTableRegionsProcedure(final MasterProcedureEnv env, -094 final RegionInfo regionToMergeA, final RegionInfo regionToMergeB) throws IOException { -095this(env, regionToMergeA, regionToMergeB, false); -096 } -097 -098 public MergeTableRegionsProcedure(final MasterProcedureEnv env, -099 final RegionInfo regionToMergeA, final RegionInfo regionToMergeB, -100 final boolean forcible) throws MergeRegionException { -101this(env, new RegionInfo[] {regionToMergeA, regionToMergeB}, forcible); -102 } -103 -104 public MergeTableRegionsProcedure(final MasterProcedureEnv env, -105 final RegionInfo[] regionsToMerge, final boolean forcible) -106 throws MergeRegionException { -107super(env); -108 -109// Check daughter regions and make sure that we have valid daughter regions -110// before doing the real work. -111checkRegionsToMerge(regionsToMerge, forcible); -112 -113// WARN: make sure there is no parent region of the two merging regions in -114// hbase:meta If exists, fixing up daughters would cause daughter regions(we -115// have merged one) online again when we restart master, so we should clear -116// the parent region to prevent the above case -117// Since HBASE-7721, we don't need fix up daughters any more. so here do nothing -118this.regionsToMerge = regionsToMerge; -119this.mergedRegion = createMergedRegionInfo(regionsToMerge); -120this.forcible = forcible; -121 } -122 -123 private static void checkRegionsToMerge(final RegionInfo[] regionsToMerge, -124 final boolean forcible) throws MergeRegionException { -125// For now, we only merge 2 regions. -126// It could be extended to more than 2
[30/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html index f0775d0..40cd159 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html @@ -943,7 +943,7 @@ 935wakeServerReportEvent(serverNode); 936 } 937 -938 public void checkOnlineRegionsReportForMeta(final ServerStateNode serverNode, +938 void checkOnlineRegionsReportForMeta(final ServerStateNode serverNode, 939 final Setbyte[] regionNames) { 940try { 941 for (byte[] regionName: regionNames) { @@ -959,7 +959,7 @@ 951final RegionStateNode regionNode = regionStates.getOrCreateRegionNode(hri); 952LOG.info("META REPORTED: " + regionNode); 953if (!reportTransition(regionNode, serverNode, TransitionCode.OPENED, 0)) { -954 LOG.warn("META REPORTED but no procedure found"); +954 LOG.warn("META REPORTED but no procedure found (complete?)"); 955 regionNode.setRegionLocation(serverNode.getServerName()); 956} else if (LOG.isTraceEnabled()) { 957 LOG.trace("META REPORTED: " + regionNode); @@ -1191,671 +1191,685 @@ 1183 public void visitRegionState(final RegionInfo regionInfo, final State state, 1184 final ServerName regionLocation, final ServerName lastHost, final long openSeqNum) { 1185final RegionStateNode regionNode = regionStates.getOrCreateRegionNode(regionInfo); -1186synchronized (regionNode) { -1187 if (!regionNode.isInTransition()) { -1188 regionNode.setState(state); -1189 regionNode.setLastHost(lastHost); -1190 regionNode.setRegionLocation(regionLocation); -1191 regionNode.setOpenSeqNum(openSeqNum); -1192 -1193if (state == State.OPEN) { -1194 assert regionLocation != null : "found null region location for " + regionNode; -1195 regionStates.addRegionToServer(regionLocation, regionNode); -1196} else if (state == State.OFFLINE || regionInfo.isOffline()) { -1197 regionStates.addToOfflineRegions(regionNode); -1198} else { -1199 // These regions should have a procedure in replay -1200 regionStates.addRegionInTransition(regionNode, null); -1201} -1202 } -1203} -1204 } -1205}); -1206 -1207// every assignment is blocked until meta is loaded. -1208wakeMetaLoadedEvent(); -1209 } -1210 -1211 // TODO: the assumption here is that if RSs are crashing while we are executing this -1212 // they will be handled by the SSH that are put in the ServerManager "queue". -1213 // we can integrate this a bit better. -1214 private boolean processofflineServersWithOnlineRegions() { -1215boolean failover = !master.getServerManager().getDeadServers().isEmpty(); -1216 -1217final SetServerName offlineServersWithOnlineRegions = new HashSetServerName(); -1218final ArrayListRegionInfo regionsToAssign = new ArrayListRegionInfo(); -1219long st, et; -1220 -1221st = System.currentTimeMillis(); -1222for (RegionStateNode regionNode: regionStates.getRegionNodes()) { -1223 if (regionNode.getState() == State.OPEN) { -1224final ServerName serverName = regionNode.getRegionLocation(); -1225if (!master.getServerManager().isServerOnline(serverName)) { -1226 offlineServersWithOnlineRegions.add(serverName); -1227} -1228 } else if (regionNode.getState() == State.OFFLINE) { -1229if (isTableEnabled(regionNode.getTable())) { -1230 regionsToAssign.add(regionNode.getRegionInfo()); -1231} -1232 } -1233} -1234et = System.currentTimeMillis(); -1235LOG.info("[STEP-1] " + StringUtils.humanTimeDiff(et - st)); -1236 -1237// kill servers with online regions -1238st = System.currentTimeMillis(); -1239for (ServerName serverName: offlineServersWithOnlineRegions) { -1240 if (!master.getServerManager().isServerOnline(serverName)) { -1241LOG.info("KILL RS hosting regions but not online " + serverName + -1242 " (master=" + master.getServerName() + ")"); -1243killRegionServer(serverName); -1244 } -1245} -1246et = System.currentTimeMillis(); -1247LOG.info("[STEP-2] " + StringUtils.humanTimeDiff(et - st)); -1248 -1249setFailoverCleanupDone(true); -1250 -1251// assign offline regions
[39/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html index cfafcf8..90e8860 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":6,"i4":10,"i5":10,"i6":6,"i7":6,"i8":10,"i9":10,"i10":10,"i11":10,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":10,"i18":6,"i19":10,"i20":10,"i21":6,"i22":10,"i23":6,"i24":10,"i25":10,"i26":6}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":6,"i4":10,"i5":10,"i6":6,"i7":6,"i8":10,"i9":10,"i10":10,"i11":10,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":10,"i18":10,"i19":6,"i20":10,"i21":10,"i22":6,"i23":10,"i24":6,"i25":10,"i26":10,"i27":6}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -162,34 +162,38 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? Field and Description +protected boolean +atomic + + protected Durability durability - + protected http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCell[] familyCellMaps - + protected int nextIndexToProcess - + protected HRegion.ObservedExceptionsInBatch observedExceptions - + protected T[] operations - + protected HRegion region - + protected OperationStatus[] retCodeDetails - + protected WALEdit[] walEditsFromCoprocessors @@ -320,25 +324,29 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getOrigLogSeqNum() +(package private) boolean +isAtomic() + + boolean isDone() - + abstract boolean isInReplay() - + boolean isOperationPending(intindex) - + MiniBatchOperationInProgressMutation lockRowsAndBuildMiniBatch(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegion.RowLockacquiredRowLocks) Creates Mini-batch of all operations [nextIndexToProcess, lastIndexExclusive) for which a row lock can be acquired. - + abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutationminiBatchOp, longtimestamp, @@ -347,15 +355,15 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? count, tags and timestamp for all cells of all operations in a mini-batch. - + int size() - + abstract void startRegionOperation() - + void visitBatchOperations(booleanpendingOnly, intlastIndexExclusive, @@ -363,12 +371,12 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? Helper method for visiting pending/ all batch operations - + protected void writeMiniBatchOperationsToMemStore(MiniBatchOperationInProgressMutationminiBatchOp, longwriteNumber) - + abstract MultiVersionConcurrencyControl.WriteEntry writeMiniBatchOperationsToMemStore(MiniBatchOperationInProgressMutationminiBatchOp, MultiVersionConcurrencyControl.WriteEntrywriteEntry) @@ -463,12 +471,21 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? - + durability protectedDurability durability + + + + + +atomic +protectedboolean atomic + + @@ -485,7 +502,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? BatchOperation -publicBatchOperation(HRegionregion, +publicBatchOperation(HRegionregion, T[]operations) @@ -503,7 +520,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? visitBatchOperations -publicvoidvisitBatchOperations(booleanpendingOnly, +publicvoidvisitBatchOperations(booleanpendingOnly, intlastIndexExclusive, HRegion.BatchOperation.Visitorvisitor) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -520,7 +537,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getMutation -public abstractMutationgetMutation(intindex) +public abstractMutationgetMutation(intindex) @@ -529,7 +546,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getNonceGroup -public
[35/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html index eaeba94..aae2da5 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html @@ -105,7 +105,7 @@ var activeTableTab = "activeTableTab"; -static interface RSRpcServices.LogDelegate +static interface RSRpcServices.LogDelegate @@ -151,7 +151,7 @@ var activeTableTab = "activeTableTab"; logBatchWarning -voidlogBatchWarning(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringfirstRegionName, +voidlogBatchWarning(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringfirstRegionName, intsum, introwSizeWarnThreshold)
[34/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html index a2d5acd..4a9326a 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html @@ -516,12 +516,13 @@ implements private void -doBatchOp(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.RegionActionResult.Builderbuilder, +doBatchOp(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.RegionActionResult.Builderbuilder, HRegionregion, OperationQuotaquota, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Actionmutations, CellScannercells, - ActivePolicyEnforcementspaceQuotaEnforcement) + ActivePolicyEnforcementspaceQuotaEnforcement, + booleanatomic) Execute a list of Put/Delete mutations. @@ -1199,7 +1200,7 @@ implements DEFAULT_LOG_DELEGATE -private staticRSRpcServices.LogDelegate DEFAULT_LOG_DELEGATE +private staticRSRpcServices.LogDelegate DEFAULT_LOG_DELEGATE @@ -1208,7 +1209,7 @@ implements ld -private finalRSRpcServices.LogDelegate ld +private finalRSRpcServices.LogDelegate ld @@ -1218,7 +1219,7 @@ implements SCANNER_ALREADY_CLOSED http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true; title="class or interface in java.lang">@Deprecated -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException SCANNER_ALREADY_CLOSED +private static finalhttp://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException SCANNER_ALREADY_CLOSED Deprecated. @@ -1236,7 +1237,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java RSRpcServices -publicRSRpcServices(HRegionServerrs) +publicRSRpcServices(HRegionServerrs) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -1250,7 +1251,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java RSRpcServices -RSRpcServices(HRegionServerrs, +RSRpcServices(HRegionServerrs, RSRpcServices.LogDelegateld) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -1396,7 +1397,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java checkAndRowMutate -privatebooleancheckAndRowMutate(HRegionregion, +privatebooleancheckAndRowMutate(HRegionregion, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Actionactions, CellScannercellScanner, byte[]row, @@ -1411,7 +1412,6 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java Parameters: cellScanner - if non-null, the mutation data -- the Cell content. -comparator - @throws IOException Throws: http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -1423,7 +1423,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java append -privateResultappend(HRegionregion, +privateResultappend(HRegionregion, OperationQuotaquota, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutationProtomutation, CellScannercellScanner, @@ -1446,7 +1446,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java increment -privateResultincrement(HRegionregion, +privateResultincrement(HRegionregion, OperationQuotaquota, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutationProtomutation, CellScannercells, @@ -1471,7 +1471,7 @@ private static finalhttp://docs.oracle.com/javase/8/docs/api/java doNonAtomicRegionMutation -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCellScannabledoNonAtomicRegionMutation(HRegionregion, +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in
[25/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html index 845f5e6..84e9e52 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html @@ -98,241 +98,248 @@ 090 @Override 091 public boolean visit(final Result r) throws IOException { 092if (r != null !r.isEmpty()) { -093 long st = System.currentTimeMillis(); -094 visitMetaEntry(visitor, r); -095 long et = System.currentTimeMillis(); -096 LOG.info("[T] LOAD META PERF " + StringUtils.humanTimeDiff(et - st)); -097} else if (isDebugEnabled) { -098 LOG.debug("NULL result from meta - ignoring but this is strange."); -099} -100return true; -101 } -102}); -103 } -104 -105 private void visitMetaEntry(final RegionStateVisitor visitor, final Result result) -106 throws IOException { -107final RegionLocations rl = MetaTableAccessor.getRegionLocations(result); -108if (rl == null) return; +093 long st = 0; +094 if (LOG.isTraceEnabled()) { +095st = System.currentTimeMillis(); +096 } +097 visitMetaEntry(visitor, r); +098 if (LOG.isTraceEnabled()) { +099long et = System.currentTimeMillis(); +100LOG.trace("[T] LOAD META PERF " + StringUtils.humanTimeDiff(et - st)); +101 } +102} else if (isDebugEnabled) { +103 LOG.debug("NULL result from meta - ignoring but this is strange."); +104} +105return true; +106 } +107}); +108 } 109 -110final HRegionLocation[] locations = rl.getRegionLocations(); -111if (locations == null) return; -112 -113for (int i = 0; i locations.length; ++i) { -114 final HRegionLocation hrl = locations[i]; -115 if (hrl == null) continue; -116 -117 final RegionInfo regionInfo = hrl.getRegionInfo(); -118 if (regionInfo == null) continue; -119 -120 final int replicaId = regionInfo.getReplicaId(); -121 final State state = getRegionState(result, replicaId); -122 -123 final ServerName lastHost = hrl.getServerName(); -124 final ServerName regionLocation = getRegionServer(result, replicaId); -125 final long openSeqNum = -1; -126 -127 // TODO: move under trace, now is visible for debugging -128 LOG.info(String.format("Load hbase:meta entry region=%s regionState=%s lastHost=%s regionLocation=%s", -129regionInfo, state, lastHost, regionLocation)); -130 -131 visitor.visitRegionState(regionInfo, state, regionLocation, lastHost, openSeqNum); -132} -133 } -134 -135 public void updateRegionLocation(final RegionInfo regionInfo, final State state, -136 final ServerName regionLocation, final ServerName lastHost, final long openSeqNum, -137 final long pid) -138 throws IOException { -139if (regionInfo.isMetaRegion()) { -140 updateMetaLocation(regionInfo, regionLocation); -141} else { -142 updateUserRegionLocation(regionInfo, state, regionLocation, lastHost, openSeqNum, pid); -143} -144 } -145 -146 public void updateRegionState(final long openSeqNum, final long pid, -147 final RegionState newState, final RegionState oldState) throws IOException { -148 updateRegionLocation(newState.getRegion(), newState.getState(), newState.getServerName(), -149oldState != null ? oldState.getServerName() : null, openSeqNum, pid); -150 } -151 -152 protected void updateMetaLocation(final RegionInfo regionInfo, final ServerName serverName) -153 throws IOException { -154try { -155 MetaTableLocator.setMetaLocation(master.getZooKeeper(), serverName, -156regionInfo.getReplicaId(), State.OPEN); -157} catch (KeeperException e) { -158 throw new IOException(e); -159} -160 } -161 -162 protected void updateUserRegionLocation(final RegionInfo regionInfo, final State state, -163 final ServerName regionLocation, final ServerName lastHost, final long openSeqNum, -164 final long pid) -165 throws IOException { -166final int replicaId = regionInfo.getReplicaId(); -167final Put put = new Put(MetaTableAccessor.getMetaKeyForRegion(regionInfo)); -168MetaTableAccessor.addRegionInfo(put, regionInfo); -169final StringBuilder info = new StringBuilder("pid=" + pid + " updating hbase:meta row="); -170 info.append(regionInfo.getRegionNameAsString()).append(", regionState=").append(state); -171if (openSeqNum = 0) { -172
[22/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html index 8e0a6a2..08916af 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateNode.html @@ -75,7 +75,7 @@ 067 }; 068 069 protected static final State[] STATES_EXPECTED_ON_CLOSE = new State[] { -070State.SPLITTING, State.SPLIT, // ServerCrashProcedure +070State.SPLITTING, State.SPLIT, State.MERGING, // ServerCrashProcedure 071State.OPEN, // enabled/open 072State.CLOSING // already in-progress (retrying) 073 }; @@ -550,446 +550,454 @@ 542return !getTableRegionStates(tableName).isEmpty(); 543 } 544 -545 public ListRegionInfo getRegionsOfTable(final TableName table) { -546return getRegionsOfTable(table, false); -547 } -548 -549 ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { -550final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); -551final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size()); -552for (RegionStateNode node: nodes) { -553 if (include(node, offline)) hris.add(node.getRegionInfo()); -554} -555return hris; -556 } -557 -558 /** -559 * Utility. Whether to include region in list of regions. Default is to -560 * weed out split and offline regions. -561 * @return True if we should include the codenode/code (do not include -562 * if split or offline unless codeoffline/code is set to true. -563 */ -564 boolean include(final RegionStateNode node, final boolean offline) { -565if (LOG.isTraceEnabled()) { -566 LOG.trace("WORKING ON " + node + " " + node.getRegionInfo()); -567} -568if (node.isInState(State.SPLIT)) return false; -569if (node.isInState(State.OFFLINE) !offline) return false; -570final RegionInfo hri = node.getRegionInfo(); -571return (!hri.isOffline() !hri.isSplit()) || -572((hri.isOffline() || hri.isSplit()) offline); -573 } -574 -575 /** -576 * Returns the set of regions hosted by the specified server -577 * @param serverName the server we are interested in -578 * @return set of RegionInfo hosted by the specified server -579 */ -580 public ListRegionInfo getServerRegionInfoSet(final ServerName serverName) { -581final ServerStateNode serverInfo = getServerNode(serverName); -582if (serverInfo == null) return Collections.emptyList(); -583 -584synchronized (serverInfo) { -585 return serverInfo.getRegionInfoList(); -586} -587 } -588 -589 // -590 // TODO: split helpers -591 // -592 public void logSplit(final ServerName serverName) { -593final ServerStateNode serverNode = getOrCreateServer(serverName); -594synchronized (serverNode) { -595 serverNode.setState(ServerState.SPLITTING); -596 /* THIS HAS TO BE WRONG. THIS IS SPLITTING OF REGION, NOT SPLITTING WALs. -597 for (RegionStateNode regionNode: serverNode.getRegions()) { -598synchronized (regionNode) { -599 // TODO: Abort procedure if present -600 regionNode.setState(State.SPLITTING); -601} -602 }*/ -603} -604 } -605 -606 public void logSplit(final RegionInfo regionInfo) { -607final RegionStateNode regionNode = getRegionNode(regionInfo); -608synchronized (regionNode) { -609 regionNode.setState(State.SPLIT); -610} -611 } -612 -613 @VisibleForTesting -614 public void updateRegionState(final RegionInfo regionInfo, final State state) { -615final RegionStateNode regionNode = getOrCreateRegionNode(regionInfo); +545 /** +546 * @return Return online regions of table; does not include OFFLINE or SPLITTING regions. +547 */ +548 public ListRegionInfo getRegionsOfTable(final TableName table) { +549return getRegionsOfTable(table, false); +550 } +551 +552 /** +553 * @return Return the regions of the table; does not include OFFLINE unless you set +554 * codeoffline/code to true. Does not include regions that are in the +555 * {@link State#SPLIT} state. +556 */ +557 public ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { +558final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); +559final ArrayListRegionInfo hris = new
[51/51] [partial] hbase-site git commit: Published site at .
Published site at . Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/a108018f Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/a108018f Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/a108018f Branch: refs/heads/asf-site Commit: a108018f14235b3904d9e73288bf8aabded259f3 Parents: d0ae29f Author: jenkinsAuthored: Fri Nov 10 15:29:29 2017 + Committer: jenkins Committed: Fri Nov 10 15:29:29 2017 + -- acid-semantics.html | 4 +- apache_hbase_reference_guide.pdf| 4 +- apidocs/allclasses-frame.html | 1 + apidocs/allclasses-noframe.html | 1 + apidocs/index-all.html |10 + apidocs/org/apache/hadoop/hbase/AuthUtil.html | 4 +- .../apache/hadoop/hbase/CacheEvictionStats.html | 269 + .../hadoop/hbase/CallDroppedException.html | 4 +- .../hbase/class-use/CacheEvictionStats.html | 169 + .../hadoop/hbase/class-use/TableName.html | 2 +- .../org/apache/hadoop/hbase/client/Admin.html | 6 +- .../org/apache/hadoop/hbase/package-frame.html | 1 + .../apache/hadoop/hbase/package-summary.html|40 +- .../org/apache/hadoop/hbase/package-tree.html | 1 + .../org/apache/hadoop/hbase/package-use.html|37 +- apidocs/overview-tree.html | 1 + .../apache/hadoop/hbase/CacheEvictionStats.html | 126 + book.html | 2 +- bulk-loads.html | 4 +- checkstyle-aggregate.html | 1470 +- checkstyle.rss |66 +- coc.html| 4 +- cygwin.html | 4 +- dependencies.html | 4 +- dependency-convergence.html |10 +- dependency-info.html| 4 +- dependency-management.html |78 +- devapidocs/allclasses-frame.html| 2 + devapidocs/allclasses-noframe.html | 2 + devapidocs/constant-values.html |58 +- devapidocs/index-all.html |74 +- .../org/apache/hadoop/hbase/Coprocessor.html| 2 +- .../hadoop/hbase/class-use/CellBuilder.html |22 + .../hadoop/hbase/class-use/CellScanner.html | 5 +- .../hadoop/hbase/class-use/Coprocessor.html |13 + .../hbase/class-use/CoprocessorEnvironment.html | 8 + .../hadoop/hbase/class-use/TableName.html | 2 +- .../hadoop/hbase/client/class-use/Mutation.html |44 +- .../hbase/client/class-use/RegionInfo.html | 2 +- .../hadoop/hbase/client/class-use/Scan.html |47 +- .../hadoop/hbase/client/package-tree.html |24 +- .../apache/hadoop/hbase/client/package-use.html | 5 + .../hbase/coprocessor/RegionCoprocessor.html| 2 +- .../hbase/coprocessor/RegionObserver.html | 2 +- .../coprocessor/class-use/ObserverContext.html |36 +- .../class-use/RegionCoprocessor.html|13 + .../class-use/RegionCoprocessorEnvironment.html |36 +- .../coprocessor/class-use/RegionObserver.html |21 + ...serverWithMetrics.ExampleRegionObserver.html | 119 +- .../ExampleRegionObserverWithMetrics.html |42 +- .../coprocessor/example/RowCountEndpoint.html | 4 +- .../example/ScanModifyingObserver.html | 477 + .../example/ValueRewritingObserver.html | 515 + .../example/WriteHeavyIncrementObserver.html| 4 +- .../class-use/ScanModifyingObserver.html| 125 + .../class-use/ValueRewritingObserver.html | 125 + .../coprocessor/example/package-frame.html | 2 + .../coprocessor/example/package-summary.html|13 + .../hbase/coprocessor/example/package-tree.html | 2 + .../hadoop/hbase/executor/package-tree.html | 2 +- .../hadoop/hbase/filter/package-tree.html |12 +- .../hadoop/hbase/io/hfile/package-tree.html | 6 +- .../apache/hadoop/hbase/ipc/package-tree.html | 2 +- .../hadoop/hbase/mapreduce/package-tree.html| 6 +- .../master/assignment/AssignmentManager.html| 100 +- .../assignment/MergeTableRegionsProcedure.html | 128 +- .../master/assignment/RegionStateStore.html |38 +- .../RegionStates.RegionFailedOpen.html |22 +- .../hbase/master/assignment/RegionStates.html |94 +- ...tTableRegionProcedure.StoreFileSplitter.html |12 +- .../assignment/SplitTableRegionProcedure.html |82 +- .../class-use/RegionStates.ServerStateNode.html | 2 +- .../hbase/master/balancer/package-tree.html | 2 +-
[40/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html index 561d099..7d7a8b0 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html @@ -134,7 +134,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class SplitTableRegionProcedure +public class SplitTableRegionProcedure extends AbstractStateMachineRegionProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionState The procedure to split a region in a table. Takes lock on the parent region. @@ -516,7 +516,7 @@ extends LOG -private static finalorg.apache.commons.logging.Log LOG +private static finalorg.apache.commons.logging.Log LOG @@ -525,7 +525,7 @@ extends traceEnabled -privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled +privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled @@ -534,7 +534,7 @@ extends daughter_1_RI -privateRegionInfo daughter_1_RI +privateRegionInfo daughter_1_RI @@ -543,7 +543,7 @@ extends daughter_2_RI -privateRegionInfo daughter_2_RI +privateRegionInfo daughter_2_RI @@ -552,7 +552,7 @@ extends bestSplitRow -privatebyte[] bestSplitRow +privatebyte[] bestSplitRow @@ -561,7 +561,7 @@ extends EXPECTED_SPLIT_STATES -private staticRegionState.State[] EXPECTED_SPLIT_STATES +private staticRegionState.State[] EXPECTED_SPLIT_STATES @@ -578,7 +578,7 @@ extends SplitTableRegionProcedure -publicSplitTableRegionProcedure() +publicSplitTableRegionProcedure() @@ -587,7 +587,7 @@ extends SplitTableRegionProcedure -publicSplitTableRegionProcedure(MasterProcedureEnvenv, +publicSplitTableRegionProcedure(MasterProcedureEnvenv, RegionInforegionToSplit, byte[]splitRow) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -611,7 +611,7 @@ extends checkSplittable -privatevoidcheckSplittable(MasterProcedureEnvenv, +privatevoidcheckSplittable(MasterProcedureEnvenv, RegionInforegionToSplit, byte[]splitRow) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -632,7 +632,7 @@ extends getDaughterRegionIdTimestamp -private staticlonggetDaughterRegionIdTimestamp(RegionInfohri) +private staticlonggetDaughterRegionIdTimestamp(RegionInfohri) Calculate daughter regionid to use. Parameters: @@ -648,7 +648,7 @@ extends executeFromState -protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv, +protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionStatestate) throws http://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException Description copied from class:StateMachineProcedure @@ -671,7 +671,7 @@ extends rollbackState -protectedvoidrollbackState(MasterProcedureEnvenv, +protectedvoidrollbackState(MasterProcedureEnvenv, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionStatestate) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException, http://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException @@ -693,7 +693,7 @@ extends isRollbackSupported -protectedbooleanisRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionStatestate) +protectedbooleanisRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionStatestate) Description copied from class:StateMachineProcedure Used by the default implementation of abort() to know if the current state can be aborted
[41/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html index e5363a7..0d3ffae 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -public static final class RegionStates.RegionFailedOpen +public static final class RegionStates.RegionFailedOpen extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -229,7 +229,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? regionNode -private finalRegionStates.RegionStateNode regionNode +private finalRegionStates.RegionStateNode regionNode @@ -238,7 +238,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? exception -private volatilehttp://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception exception +private volatilehttp://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception exception @@ -247,7 +247,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? retries -private volatileint retries +private volatileint retries @@ -264,7 +264,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? RegionFailedOpen -publicRegionFailedOpen(RegionStates.RegionStateNoderegionNode) +publicRegionFailedOpen(RegionStates.RegionStateNoderegionNode) @@ -281,7 +281,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getRegionNode -publicRegionStates.RegionStateNodegetRegionNode() +publicRegionStates.RegionStateNodegetRegionNode() @@ -290,7 +290,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getRegionInfo -publicRegionInfogetRegionInfo() +publicRegionInfogetRegionInfo() @@ -299,7 +299,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? incrementAndGetRetries -publicintincrementAndGetRetries() +publicintincrementAndGetRetries() @@ -308,7 +308,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getRetries -publicintgetRetries() +publicintgetRetries() @@ -317,7 +317,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? setException -publicvoidsetException(http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exceptionexception) +publicvoidsetException(http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exceptionexception) @@ -326,7 +326,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getException -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">ExceptiongetException() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">ExceptiongetException() http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html index c71b9c3..e81ea36 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html @@ -365,7 +365,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getRegionsOfTable(TableNametable) -(package private) http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo getRegionsOfTable(TableNametable, booleanoffline) @@ -784,7 +784,11 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getRegionsOfTable -publichttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in
[28/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html index f0775d0..40cd159 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html @@ -943,7 +943,7 @@ 935wakeServerReportEvent(serverNode); 936 } 937 -938 public void checkOnlineRegionsReportForMeta(final ServerStateNode serverNode, +938 void checkOnlineRegionsReportForMeta(final ServerStateNode serverNode, 939 final Setbyte[] regionNames) { 940try { 941 for (byte[] regionName: regionNames) { @@ -959,7 +959,7 @@ 951final RegionStateNode regionNode = regionStates.getOrCreateRegionNode(hri); 952LOG.info("META REPORTED: " + regionNode); 953if (!reportTransition(regionNode, serverNode, TransitionCode.OPENED, 0)) { -954 LOG.warn("META REPORTED but no procedure found"); +954 LOG.warn("META REPORTED but no procedure found (complete?)"); 955 regionNode.setRegionLocation(serverNode.getServerName()); 956} else if (LOG.isTraceEnabled()) { 957 LOG.trace("META REPORTED: " + regionNode); @@ -1191,671 +1191,685 @@ 1183 public void visitRegionState(final RegionInfo regionInfo, final State state, 1184 final ServerName regionLocation, final ServerName lastHost, final long openSeqNum) { 1185final RegionStateNode regionNode = regionStates.getOrCreateRegionNode(regionInfo); -1186synchronized (regionNode) { -1187 if (!regionNode.isInTransition()) { -1188 regionNode.setState(state); -1189 regionNode.setLastHost(lastHost); -1190 regionNode.setRegionLocation(regionLocation); -1191 regionNode.setOpenSeqNum(openSeqNum); -1192 -1193if (state == State.OPEN) { -1194 assert regionLocation != null : "found null region location for " + regionNode; -1195 regionStates.addRegionToServer(regionLocation, regionNode); -1196} else if (state == State.OFFLINE || regionInfo.isOffline()) { -1197 regionStates.addToOfflineRegions(regionNode); -1198} else { -1199 // These regions should have a procedure in replay -1200 regionStates.addRegionInTransition(regionNode, null); -1201} -1202 } -1203} -1204 } -1205}); -1206 -1207// every assignment is blocked until meta is loaded. -1208wakeMetaLoadedEvent(); -1209 } -1210 -1211 // TODO: the assumption here is that if RSs are crashing while we are executing this -1212 // they will be handled by the SSH that are put in the ServerManager "queue". -1213 // we can integrate this a bit better. -1214 private boolean processofflineServersWithOnlineRegions() { -1215boolean failover = !master.getServerManager().getDeadServers().isEmpty(); -1216 -1217final SetServerName offlineServersWithOnlineRegions = new HashSetServerName(); -1218final ArrayListRegionInfo regionsToAssign = new ArrayListRegionInfo(); -1219long st, et; -1220 -1221st = System.currentTimeMillis(); -1222for (RegionStateNode regionNode: regionStates.getRegionNodes()) { -1223 if (regionNode.getState() == State.OPEN) { -1224final ServerName serverName = regionNode.getRegionLocation(); -1225if (!master.getServerManager().isServerOnline(serverName)) { -1226 offlineServersWithOnlineRegions.add(serverName); -1227} -1228 } else if (regionNode.getState() == State.OFFLINE) { -1229if (isTableEnabled(regionNode.getTable())) { -1230 regionsToAssign.add(regionNode.getRegionInfo()); -1231} -1232 } -1233} -1234et = System.currentTimeMillis(); -1235LOG.info("[STEP-1] " + StringUtils.humanTimeDiff(et - st)); -1236 -1237// kill servers with online regions -1238st = System.currentTimeMillis(); -1239for (ServerName serverName: offlineServersWithOnlineRegions) { -1240 if (!master.getServerManager().isServerOnline(serverName)) { -1241LOG.info("KILL RS hosting regions but not online " + serverName + -1242 " (master=" + master.getServerName() + ")"); -1243killRegionServer(serverName); -1244 } -1245} -1246et = System.currentTimeMillis(); -1247LOG.info("[STEP-2] " + StringUtils.humanTimeDiff(et - st)); -1248 -1249setFailoverCleanupDone(true); -1250 -1251// assign offline regions -1252st = System.currentTimeMillis(); -1253for (RegionInfo regionInfo: getOrderedRegions(regionsToAssign)) {
[16/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html index e844201..496c92f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.StoreFileSplitter.html @@ -50,768 +50,774 @@ 042import org.apache.hadoop.hbase.HConstants; 043import org.apache.hadoop.hbase.ServerName; 044import org.apache.hadoop.hbase.TableName; -045import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -046import org.apache.hadoop.hbase.client.MasterSwitchType; -047import org.apache.hadoop.hbase.client.Mutation; -048import org.apache.hadoop.hbase.client.RegionInfo; -049import org.apache.hadoop.hbase.client.RegionInfoBuilder; -050import org.apache.hadoop.hbase.client.RegionReplicaUtil; -051import org.apache.hadoop.hbase.client.TableDescriptor; -052import org.apache.hadoop.hbase.io.hfile.CacheConfig; -053import org.apache.hadoop.hbase.master.MasterCoprocessorHost; -054import org.apache.hadoop.hbase.master.MasterFileSystem; -055import org.apache.hadoop.hbase.master.RegionState.State; -056import org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode; -057import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -058import org.apache.hadoop.hbase.master.procedure.AbstractStateMachineRegionProcedure; -059import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; -060import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil; -061import org.apache.hadoop.hbase.procedure2.ProcedureMetrics; -062import org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer; -063import org.apache.hadoop.hbase.quotas.QuotaExceededException; -064import org.apache.hadoop.hbase.regionserver.HRegionFileSystem; -065import org.apache.hadoop.hbase.regionserver.HStore; -066import org.apache.hadoop.hbase.regionserver.HStoreFile; -067import org.apache.hadoop.hbase.regionserver.StoreFileInfo; -068import org.apache.hadoop.hbase.util.Bytes; -069import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -070import org.apache.hadoop.hbase.util.FSUtils; -071import org.apache.hadoop.hbase.util.Pair; -072import org.apache.hadoop.hbase.util.Threads; -073import org.apache.yetus.audience.InterfaceAudience; -074 -075import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting; -076import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -077import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -078import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos; -079import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionState; -080 -081/** -082 * The procedure to split a region in a table. -083 * Takes lock on the parent region. -084 * It holds the lock for the life of the procedure. -085 */ -086@InterfaceAudience.Private -087public class SplitTableRegionProcedure -088extends AbstractStateMachineRegionProcedureSplitTableRegionState { -089 private static final Log LOG = LogFactory.getLog(SplitTableRegionProcedure.class); -090 private Boolean traceEnabled = null; -091 private RegionInfo daughter_1_RI; -092 private RegionInfo daughter_2_RI; -093 private byte[] bestSplitRow; -094 -095 public SplitTableRegionProcedure() { -096// Required by the Procedure framework to create the procedure on replay -097 } -098 -099 public SplitTableRegionProcedure(final MasterProcedureEnv env, -100 final RegionInfo regionToSplit, final byte[] splitRow) throws IOException { -101super(env, regionToSplit); -102this.bestSplitRow = splitRow; -103checkSplittable(env, regionToSplit, bestSplitRow); -104final TableName table = regionToSplit.getTable(); -105final long rid = getDaughterRegionIdTimestamp(regionToSplit); -106this.daughter_1_RI = RegionInfoBuilder.newBuilder(table) -107 .setStartKey(regionToSplit.getStartKey()) -108.setEndKey(bestSplitRow) -109.setSplit(false) -110.setRegionId(rid) -111.build(); -112this.daughter_2_RI = RegionInfoBuilder.newBuilder(table) -113.setStartKey(bestSplitRow) -114 .setEndKey(regionToSplit.getEndKey()) -115.setSplit(false) -116.setRegionId(rid) -117.build(); -118 } -119 -120 /** -121 * Check whether the region is splittable -122 * @param env MasterProcedureEnv -123 * @param regionToSplit parent Region to be split -124 * @param
[05/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode, FSWALEntry#stampRegionSequenceId
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html index 7a70a5b..a7535b9 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ExampleRegionObserverWithMetrics.ExampleRegionObserver.html @@ -43,99 +43,141 @@ 035import org.apache.hadoop.hbase.metrics.Counter; 036import org.apache.hadoop.hbase.metrics.MetricRegistry; 037import org.apache.hadoop.hbase.metrics.Timer; -038 -039/** -040 * An example coprocessor that collects some metrics to demonstrate the usage of exporting custom -041 * metrics from the coprocessor. -042 * p -043 * These metrics will be available through the regular Hadoop metrics2 sinks (ganglia, opentsdb, -044 * etc) as well as JMX output. You can view a snapshot of the metrics by going to the http web UI -045 * of the regionserver page, something like http://myregionserverhost:16030/jmx -046 * /p -047 * -048 * @see ExampleMasterObserverWithMetrics -049 */ -050public class ExampleRegionObserverWithMetrics implements RegionCoprocessor { -051 -052 private Counter preGetCounter; -053 private Timer costlyOperationTimer; -054 private ExampleRegionObserver observer; -055 -056 class ExampleRegionObserver implements RegionCoprocessor, RegionObserver { -057@Override -058public OptionalRegionObserver getRegionObserver() { -059 return Optional.of(this); -060} -061 -062@Override -063public void preGetOp(ObserverContextRegionCoprocessorEnvironment e, Get get, -064ListCell results) throws IOException { -065 // Increment the Counter whenever the coprocessor is called -066 preGetCounter.increment(); +038import org.apache.hadoop.hbase.regionserver.FlushLifeCycleTracker; +039import org.apache.hadoop.hbase.regionserver.Store; +040import org.apache.hadoop.hbase.regionserver.StoreFile; +041import org.apache.hadoop.hbase.regionserver.compactions.CompactionLifeCycleTracker; +042import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest; +043 +044/** +045 * An example coprocessor that collects some metrics to demonstrate the usage of exporting custom +046 * metrics from the coprocessor. +047 * p +048 * These metrics will be available through the regular Hadoop metrics2 sinks (ganglia, opentsdb, +049 * etc) as well as JMX output. You can view a snapshot of the metrics by going to the http web UI +050 * of the regionserver page, something like http://myregionserverhost:16030/jmx +051 * /p +052 * +053 * @see ExampleMasterObserverWithMetrics +054 */ +055public class ExampleRegionObserverWithMetrics implements RegionCoprocessor { +056 +057 private Counter preGetCounter; +058 private Counter flushCounter; +059 private Counter filesCompactedCounter; +060 private Timer costlyOperationTimer; +061 private ExampleRegionObserver observer; +062 +063 class ExampleRegionObserver implements RegionCoprocessor, RegionObserver { +064@Override +065public OptionalRegionObserver getRegionObserver() { +066 return Optional.of(this); 067} 068 069@Override -070public void postGetOp(ObserverContextRegionCoprocessorEnvironment e, Get get, +070public void preGetOp(ObserverContextRegionCoprocessorEnvironment e, Get get, 071ListCell results) throws IOException { -072 // do a costly (high latency) operation which we want to measure how long it takes by -073 // using a Timer (which is a Meter and a Histogram). -074 long start = System.nanoTime(); -075 try { -076performCostlyOperation(); -077 } finally { -078 costlyOperationTimer.updateNanos(System.nanoTime() - start); -079 } -080} -081 -082private void performCostlyOperation() { -083 try { -084// simulate the operation by sleeping. -085 Thread.sleep(ThreadLocalRandom.current().nextLong(100)); -086 } catch (InterruptedException ignore) { -087 } -088} -089 } -090 -091 @Override public OptionalRegionObserver getRegionObserver() { -092return Optional.of(observer); -093 } -094 -095 @Override -096 public void start(CoprocessorEnvironment env) throws IOException { -097// start for the RegionServerObserver will be called only once in the lifetime of the -098// server. We will construct and register all metrics that we will track across method -099// invocations. -100 -101if (env instanceof RegionCoprocessorEnvironment) {
[07/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode,
[17/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html index 8e0a6a2..08916af 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html @@ -75,7 +75,7 @@ 067 }; 068 069 protected static final State[] STATES_EXPECTED_ON_CLOSE = new State[] { -070State.SPLITTING, State.SPLIT, // ServerCrashProcedure +070State.SPLITTING, State.SPLIT, State.MERGING, // ServerCrashProcedure 071State.OPEN, // enabled/open 072State.CLOSING // already in-progress (retrying) 073 }; @@ -550,446 +550,454 @@ 542return !getTableRegionStates(tableName).isEmpty(); 543 } 544 -545 public ListRegionInfo getRegionsOfTable(final TableName table) { -546return getRegionsOfTable(table, false); -547 } -548 -549 ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { -550final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); -551final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size()); -552for (RegionStateNode node: nodes) { -553 if (include(node, offline)) hris.add(node.getRegionInfo()); -554} -555return hris; -556 } -557 -558 /** -559 * Utility. Whether to include region in list of regions. Default is to -560 * weed out split and offline regions. -561 * @return True if we should include the codenode/code (do not include -562 * if split or offline unless codeoffline/code is set to true. -563 */ -564 boolean include(final RegionStateNode node, final boolean offline) { -565if (LOG.isTraceEnabled()) { -566 LOG.trace("WORKING ON " + node + " " + node.getRegionInfo()); -567} -568if (node.isInState(State.SPLIT)) return false; -569if (node.isInState(State.OFFLINE) !offline) return false; -570final RegionInfo hri = node.getRegionInfo(); -571return (!hri.isOffline() !hri.isSplit()) || -572((hri.isOffline() || hri.isSplit()) offline); -573 } -574 -575 /** -576 * Returns the set of regions hosted by the specified server -577 * @param serverName the server we are interested in -578 * @return set of RegionInfo hosted by the specified server -579 */ -580 public ListRegionInfo getServerRegionInfoSet(final ServerName serverName) { -581final ServerStateNode serverInfo = getServerNode(serverName); -582if (serverInfo == null) return Collections.emptyList(); -583 -584synchronized (serverInfo) { -585 return serverInfo.getRegionInfoList(); -586} -587 } -588 -589 // -590 // TODO: split helpers -591 // -592 public void logSplit(final ServerName serverName) { -593final ServerStateNode serverNode = getOrCreateServer(serverName); -594synchronized (serverNode) { -595 serverNode.setState(ServerState.SPLITTING); -596 /* THIS HAS TO BE WRONG. THIS IS SPLITTING OF REGION, NOT SPLITTING WALs. -597 for (RegionStateNode regionNode: serverNode.getRegions()) { -598synchronized (regionNode) { -599 // TODO: Abort procedure if present -600 regionNode.setState(State.SPLITTING); -601} -602 }*/ -603} -604 } -605 -606 public void logSplit(final RegionInfo regionInfo) { -607final RegionStateNode regionNode = getRegionNode(regionInfo); -608synchronized (regionNode) { -609 regionNode.setState(State.SPLIT); -610} -611 } -612 -613 @VisibleForTesting -614 public void updateRegionState(final RegionInfo regionInfo, final State state) { -615final RegionStateNode regionNode = getOrCreateRegionNode(regionInfo); +545 /** +546 * @return Return online regions of table; does not include OFFLINE or SPLITTING regions. +547 */ +548 public ListRegionInfo getRegionsOfTable(final TableName table) { +549return getRegionsOfTable(table, false); +550 } +551 +552 /** +553 * @return Return the regions of the table; does not include OFFLINE unless you set +554 * codeoffline/code to true. Does not include regions that are in the +555 * {@link State#SPLIT} state. +556 */ +557 public ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { +558final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); +559final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size()); +560for (RegionStateNode node: nodes) { +561 if
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html index 8e0a6a2..08916af 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html @@ -75,7 +75,7 @@ 067 }; 068 069 protected static final State[] STATES_EXPECTED_ON_CLOSE = new State[] { -070State.SPLITTING, State.SPLIT, // ServerCrashProcedure +070State.SPLITTING, State.SPLIT, State.MERGING, // ServerCrashProcedure 071State.OPEN, // enabled/open 072State.CLOSING // already in-progress (retrying) 073 }; @@ -550,446 +550,454 @@ 542return !getTableRegionStates(tableName).isEmpty(); 543 } 544 -545 public ListRegionInfo getRegionsOfTable(final TableName table) { -546return getRegionsOfTable(table, false); -547 } -548 -549 ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { -550final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); -551final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size()); -552for (RegionStateNode node: nodes) { -553 if (include(node, offline)) hris.add(node.getRegionInfo()); -554} -555return hris; -556 } -557 -558 /** -559 * Utility. Whether to include region in list of regions. Default is to -560 * weed out split and offline regions. -561 * @return True if we should include the codenode/code (do not include -562 * if split or offline unless codeoffline/code is set to true. -563 */ -564 boolean include(final RegionStateNode node, final boolean offline) { -565if (LOG.isTraceEnabled()) { -566 LOG.trace("WORKING ON " + node + " " + node.getRegionInfo()); -567} -568if (node.isInState(State.SPLIT)) return false; -569if (node.isInState(State.OFFLINE) !offline) return false; -570final RegionInfo hri = node.getRegionInfo(); -571return (!hri.isOffline() !hri.isSplit()) || -572((hri.isOffline() || hri.isSplit()) offline); -573 } -574 -575 /** -576 * Returns the set of regions hosted by the specified server -577 * @param serverName the server we are interested in -578 * @return set of RegionInfo hosted by the specified server -579 */ -580 public ListRegionInfo getServerRegionInfoSet(final ServerName serverName) { -581final ServerStateNode serverInfo = getServerNode(serverName); -582if (serverInfo == null) return Collections.emptyList(); -583 -584synchronized (serverInfo) { -585 return serverInfo.getRegionInfoList(); -586} -587 } -588 -589 // -590 // TODO: split helpers -591 // -592 public void logSplit(final ServerName serverName) { -593final ServerStateNode serverNode = getOrCreateServer(serverName); -594synchronized (serverNode) { -595 serverNode.setState(ServerState.SPLITTING); -596 /* THIS HAS TO BE WRONG. THIS IS SPLITTING OF REGION, NOT SPLITTING WALs. -597 for (RegionStateNode regionNode: serverNode.getRegions()) { -598synchronized (regionNode) { -599 // TODO: Abort procedure if present -600 regionNode.setState(State.SPLITTING); -601} -602 }*/ -603} -604 } -605 -606 public void logSplit(final RegionInfo regionInfo) { -607final RegionStateNode regionNode = getRegionNode(regionInfo); -608synchronized (regionNode) { -609 regionNode.setState(State.SPLIT); -610} -611 } -612 -613 @VisibleForTesting -614 public void updateRegionState(final RegionInfo regionInfo, final State state) { -615final RegionStateNode regionNode = getOrCreateRegionNode(regionInfo); +545 /** +546 * @return Return online regions of table; does not include OFFLINE or SPLITTING regions. +547 */ +548 public ListRegionInfo getRegionsOfTable(final TableName table) { +549return getRegionsOfTable(table, false); +550 } +551 +552 /** +553 * @return Return the regions of the table; does not include OFFLINE unless you set +554 * codeoffline/code to true. Does not include regions that are in the +555 * {@link State#SPLIT} state. +556 */ +557 public ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { +558final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); +559final ArrayListRegionInfo hris = new
[18/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html index 8e0a6a2..08916af 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerStateNode.html @@ -75,7 +75,7 @@ 067 }; 068 069 protected static final State[] STATES_EXPECTED_ON_CLOSE = new State[] { -070State.SPLITTING, State.SPLIT, // ServerCrashProcedure +070State.SPLITTING, State.SPLIT, State.MERGING, // ServerCrashProcedure 071State.OPEN, // enabled/open 072State.CLOSING // already in-progress (retrying) 073 }; @@ -550,446 +550,454 @@ 542return !getTableRegionStates(tableName).isEmpty(); 543 } 544 -545 public ListRegionInfo getRegionsOfTable(final TableName table) { -546return getRegionsOfTable(table, false); -547 } -548 -549 ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { -550final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); -551final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size()); -552for (RegionStateNode node: nodes) { -553 if (include(node, offline)) hris.add(node.getRegionInfo()); -554} -555return hris; -556 } -557 -558 /** -559 * Utility. Whether to include region in list of regions. Default is to -560 * weed out split and offline regions. -561 * @return True if we should include the codenode/code (do not include -562 * if split or offline unless codeoffline/code is set to true. -563 */ -564 boolean include(final RegionStateNode node, final boolean offline) { -565if (LOG.isTraceEnabled()) { -566 LOG.trace("WORKING ON " + node + " " + node.getRegionInfo()); -567} -568if (node.isInState(State.SPLIT)) return false; -569if (node.isInState(State.OFFLINE) !offline) return false; -570final RegionInfo hri = node.getRegionInfo(); -571return (!hri.isOffline() !hri.isSplit()) || -572((hri.isOffline() || hri.isSplit()) offline); -573 } -574 -575 /** -576 * Returns the set of regions hosted by the specified server -577 * @param serverName the server we are interested in -578 * @return set of RegionInfo hosted by the specified server -579 */ -580 public ListRegionInfo getServerRegionInfoSet(final ServerName serverName) { -581final ServerStateNode serverInfo = getServerNode(serverName); -582if (serverInfo == null) return Collections.emptyList(); -583 -584synchronized (serverInfo) { -585 return serverInfo.getRegionInfoList(); -586} -587 } -588 -589 // -590 // TODO: split helpers -591 // -592 public void logSplit(final ServerName serverName) { -593final ServerStateNode serverNode = getOrCreateServer(serverName); -594synchronized (serverNode) { -595 serverNode.setState(ServerState.SPLITTING); -596 /* THIS HAS TO BE WRONG. THIS IS SPLITTING OF REGION, NOT SPLITTING WALs. -597 for (RegionStateNode regionNode: serverNode.getRegions()) { -598synchronized (regionNode) { -599 // TODO: Abort procedure if present -600 regionNode.setState(State.SPLITTING); -601} -602 }*/ -603} -604 } -605 -606 public void logSplit(final RegionInfo regionInfo) { -607final RegionStateNode regionNode = getRegionNode(regionInfo); -608synchronized (regionNode) { -609 regionNode.setState(State.SPLIT); -610} -611 } -612 -613 @VisibleForTesting -614 public void updateRegionState(final RegionInfo regionInfo, final State state) { -615final RegionStateNode regionNode = getOrCreateRegionNode(regionInfo); +545 /** +546 * @return Return online regions of table; does not include OFFLINE or SPLITTING regions. +547 */ +548 public ListRegionInfo getRegionsOfTable(final TableName table) { +549return getRegionsOfTable(table, false); +550 } +551 +552 /** +553 * @return Return the regions of the table; does not include OFFLINE unless you set +554 * codeoffline/code to true. Does not include regions that are in the +555 * {@link State#SPLIT} state. +556 */ +557 public ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { +558final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); +559final ArrayListRegionInfo hris = new
[33/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html index 73dd853..3dfd003 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/FlushLifeCycleTracker.html @@ -176,6 +176,18 @@ +void +ExampleRegionObserverWithMetrics.ExampleRegionObserver.postFlush(ObserverContextRegionCoprocessorEnvironmentc, + FlushLifeCycleTrackertracker) + + +void +ExampleRegionObserverWithMetrics.ExampleRegionObserver.postFlush(ObserverContextRegionCoprocessorEnvironmentc, + Storestore, + StoreFileresultFile, + FlushLifeCycleTrackertracker) + + InternalScanner WriteHeavyIncrementObserver.preFlush(ObserverContextRegionCoprocessorEnvironmentc, Storestore, http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html index 5ba23fc..ad1c278 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.BatchOperation.html @@ -103,13 +103,13 @@ -private static class +(package private) static class HRegion.MutationBatchOperation Batch of mutation operations. -private static class +(package private) static class HRegion.ReplayBatchOperation Batch of mutations for replay. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html index f872206..bfc586b 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html @@ -674,12 +674,13 @@ private void -RSRpcServices.doBatchOp(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.RegionActionResult.Builderbuilder, +RSRpcServices.doBatchOp(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.RegionActionResult.Builderbuilder, HRegionregion, OperationQuotaquota, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Actionmutations, CellScannercells, - ActivePolicyEnforcementspaceQuotaEnforcement) + ActivePolicyEnforcementspaceQuotaEnforcement, + booleanatomic) Execute a list of Put/Delete mutations. @@ -1051,8 +1052,9 @@ MetricsRegionWrapperImpl(HRegionregion) -MutationBatchOperation(HRegionregion, +MutationBatchOperation(HRegionregion, Mutation[]operations, + booleanatomic, longnonceGroup, longnonce) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html index 2952f9f..d6acc81 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/InternalScanner.html @@ -336,18 +336,27 @@ InternalScanner +ValueRewritingObserver.preCompact(ObserverContextRegionCoprocessorEnvironmentc, + Storestore, + InternalScannerscanner, + ScanTypescanType, + CompactionLifeCycleTrackertracker, + CompactionRequestrequest) + + +InternalScanner WriteHeavyIncrementObserver.preFlush(ObserverContextRegionCoprocessorEnvironmentc, Storestore, InternalScannerscanner, FlushLifeCycleTrackertracker) - + InternalScanner WriteHeavyIncrementObserver.preMemStoreCompactionCompact(ObserverContextRegionCoprocessorEnvironmentc, Storestore, InternalScannerscanner) - + private
[24/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html index 8e0a6a2..08916af 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html @@ -75,7 +75,7 @@ 067 }; 068 069 protected static final State[] STATES_EXPECTED_ON_CLOSE = new State[] { -070State.SPLITTING, State.SPLIT, // ServerCrashProcedure +070State.SPLITTING, State.SPLIT, State.MERGING, // ServerCrashProcedure 071State.OPEN, // enabled/open 072State.CLOSING // already in-progress (retrying) 073 }; @@ -550,446 +550,454 @@ 542return !getTableRegionStates(tableName).isEmpty(); 543 } 544 -545 public ListRegionInfo getRegionsOfTable(final TableName table) { -546return getRegionsOfTable(table, false); -547 } -548 -549 ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { -550final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); -551final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size()); -552for (RegionStateNode node: nodes) { -553 if (include(node, offline)) hris.add(node.getRegionInfo()); -554} -555return hris; -556 } -557 -558 /** -559 * Utility. Whether to include region in list of regions. Default is to -560 * weed out split and offline regions. -561 * @return True if we should include the codenode/code (do not include -562 * if split or offline unless codeoffline/code is set to true. -563 */ -564 boolean include(final RegionStateNode node, final boolean offline) { -565if (LOG.isTraceEnabled()) { -566 LOG.trace("WORKING ON " + node + " " + node.getRegionInfo()); -567} -568if (node.isInState(State.SPLIT)) return false; -569if (node.isInState(State.OFFLINE) !offline) return false; -570final RegionInfo hri = node.getRegionInfo(); -571return (!hri.isOffline() !hri.isSplit()) || -572((hri.isOffline() || hri.isSplit()) offline); -573 } -574 -575 /** -576 * Returns the set of regions hosted by the specified server -577 * @param serverName the server we are interested in -578 * @return set of RegionInfo hosted by the specified server -579 */ -580 public ListRegionInfo getServerRegionInfoSet(final ServerName serverName) { -581final ServerStateNode serverInfo = getServerNode(serverName); -582if (serverInfo == null) return Collections.emptyList(); -583 -584synchronized (serverInfo) { -585 return serverInfo.getRegionInfoList(); -586} -587 } -588 -589 // -590 // TODO: split helpers -591 // -592 public void logSplit(final ServerName serverName) { -593final ServerStateNode serverNode = getOrCreateServer(serverName); -594synchronized (serverNode) { -595 serverNode.setState(ServerState.SPLITTING); -596 /* THIS HAS TO BE WRONG. THIS IS SPLITTING OF REGION, NOT SPLITTING WALs. -597 for (RegionStateNode regionNode: serverNode.getRegions()) { -598synchronized (regionNode) { -599 // TODO: Abort procedure if present -600 regionNode.setState(State.SPLITTING); -601} -602 }*/ -603} -604 } -605 -606 public void logSplit(final RegionInfo regionInfo) { -607final RegionStateNode regionNode = getRegionNode(regionInfo); -608synchronized (regionNode) { -609 regionNode.setState(State.SPLIT); -610} -611 } -612 -613 @VisibleForTesting -614 public void updateRegionState(final RegionInfo regionInfo, final State state) { -615final RegionStateNode regionNode = getOrCreateRegionNode(regionInfo); +545 /** +546 * @return Return online regions of table; does not include OFFLINE or SPLITTING regions. +547 */ +548 public ListRegionInfo getRegionsOfTable(final TableName table) { +549return getRegionsOfTable(table, false); +550 } +551 +552 /** +553 * @return Return the regions of the table; does not include OFFLINE unless you set +554 * codeoffline/code to true. Does not include regions that are in the +555 * {@link State#SPLIT} state. +556 */ +557 public ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { +558final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); +559final
[15/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html index e844201..496c92f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/SplitTableRegionProcedure.html @@ -50,768 +50,774 @@ 042import org.apache.hadoop.hbase.HConstants; 043import org.apache.hadoop.hbase.ServerName; 044import org.apache.hadoop.hbase.TableName; -045import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -046import org.apache.hadoop.hbase.client.MasterSwitchType; -047import org.apache.hadoop.hbase.client.Mutation; -048import org.apache.hadoop.hbase.client.RegionInfo; -049import org.apache.hadoop.hbase.client.RegionInfoBuilder; -050import org.apache.hadoop.hbase.client.RegionReplicaUtil; -051import org.apache.hadoop.hbase.client.TableDescriptor; -052import org.apache.hadoop.hbase.io.hfile.CacheConfig; -053import org.apache.hadoop.hbase.master.MasterCoprocessorHost; -054import org.apache.hadoop.hbase.master.MasterFileSystem; -055import org.apache.hadoop.hbase.master.RegionState.State; -056import org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode; -057import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -058import org.apache.hadoop.hbase.master.procedure.AbstractStateMachineRegionProcedure; -059import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; -060import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil; -061import org.apache.hadoop.hbase.procedure2.ProcedureMetrics; -062import org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer; -063import org.apache.hadoop.hbase.quotas.QuotaExceededException; -064import org.apache.hadoop.hbase.regionserver.HRegionFileSystem; -065import org.apache.hadoop.hbase.regionserver.HStore; -066import org.apache.hadoop.hbase.regionserver.HStoreFile; -067import org.apache.hadoop.hbase.regionserver.StoreFileInfo; -068import org.apache.hadoop.hbase.util.Bytes; -069import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -070import org.apache.hadoop.hbase.util.FSUtils; -071import org.apache.hadoop.hbase.util.Pair; -072import org.apache.hadoop.hbase.util.Threads; -073import org.apache.yetus.audience.InterfaceAudience; -074 -075import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting; -076import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -077import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -078import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos; -079import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SplitTableRegionState; -080 -081/** -082 * The procedure to split a region in a table. -083 * Takes lock on the parent region. -084 * It holds the lock for the life of the procedure. -085 */ -086@InterfaceAudience.Private -087public class SplitTableRegionProcedure -088extends AbstractStateMachineRegionProcedureSplitTableRegionState { -089 private static final Log LOG = LogFactory.getLog(SplitTableRegionProcedure.class); -090 private Boolean traceEnabled = null; -091 private RegionInfo daughter_1_RI; -092 private RegionInfo daughter_2_RI; -093 private byte[] bestSplitRow; -094 -095 public SplitTableRegionProcedure() { -096// Required by the Procedure framework to create the procedure on replay -097 } -098 -099 public SplitTableRegionProcedure(final MasterProcedureEnv env, -100 final RegionInfo regionToSplit, final byte[] splitRow) throws IOException { -101super(env, regionToSplit); -102this.bestSplitRow = splitRow; -103checkSplittable(env, regionToSplit, bestSplitRow); -104final TableName table = regionToSplit.getTable(); -105final long rid = getDaughterRegionIdTimestamp(regionToSplit); -106this.daughter_1_RI = RegionInfoBuilder.newBuilder(table) -107 .setStartKey(regionToSplit.getStartKey()) -108.setEndKey(bestSplitRow) -109.setSplit(false) -110.setRegionId(rid) -111.build(); -112this.daughter_2_RI = RegionInfoBuilder.newBuilder(table) -113.setStartKey(bestSplitRow) -114 .setEndKey(regionToSplit.getEndKey()) -115.setSplit(false) -116.setRegionId(rid) -117.build(); -118 } -119 -120 /** -121 * Check whether the region is splittable -122 * @param env MasterProcedureEnv -123 * @param regionToSplit parent Region to be split -124 * @param splitRow if splitRow is not specified, will first try to get bestSplitRow from RS -125 *
[29/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html index f0775d0..40cd159 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html @@ -943,7 +943,7 @@ 935wakeServerReportEvent(serverNode); 936 } 937 -938 public void checkOnlineRegionsReportForMeta(final ServerStateNode serverNode, +938 void checkOnlineRegionsReportForMeta(final ServerStateNode serverNode, 939 final Setbyte[] regionNames) { 940try { 941 for (byte[] regionName: regionNames) { @@ -959,7 +959,7 @@ 951final RegionStateNode regionNode = regionStates.getOrCreateRegionNode(hri); 952LOG.info("META REPORTED: " + regionNode); 953if (!reportTransition(regionNode, serverNode, TransitionCode.OPENED, 0)) { -954 LOG.warn("META REPORTED but no procedure found"); +954 LOG.warn("META REPORTED but no procedure found (complete?)"); 955 regionNode.setRegionLocation(serverNode.getServerName()); 956} else if (LOG.isTraceEnabled()) { 957 LOG.trace("META REPORTED: " + regionNode); @@ -1191,671 +1191,685 @@ 1183 public void visitRegionState(final RegionInfo regionInfo, final State state, 1184 final ServerName regionLocation, final ServerName lastHost, final long openSeqNum) { 1185final RegionStateNode regionNode = regionStates.getOrCreateRegionNode(regionInfo); -1186synchronized (regionNode) { -1187 if (!regionNode.isInTransition()) { -1188 regionNode.setState(state); -1189 regionNode.setLastHost(lastHost); -1190 regionNode.setRegionLocation(regionLocation); -1191 regionNode.setOpenSeqNum(openSeqNum); -1192 -1193if (state == State.OPEN) { -1194 assert regionLocation != null : "found null region location for " + regionNode; -1195 regionStates.addRegionToServer(regionLocation, regionNode); -1196} else if (state == State.OFFLINE || regionInfo.isOffline()) { -1197 regionStates.addToOfflineRegions(regionNode); -1198} else { -1199 // These regions should have a procedure in replay -1200 regionStates.addRegionInTransition(regionNode, null); -1201} -1202 } -1203} -1204 } -1205}); -1206 -1207// every assignment is blocked until meta is loaded. -1208wakeMetaLoadedEvent(); -1209 } -1210 -1211 // TODO: the assumption here is that if RSs are crashing while we are executing this -1212 // they will be handled by the SSH that are put in the ServerManager "queue". -1213 // we can integrate this a bit better. -1214 private boolean processofflineServersWithOnlineRegions() { -1215boolean failover = !master.getServerManager().getDeadServers().isEmpty(); -1216 -1217final SetServerName offlineServersWithOnlineRegions = new HashSetServerName(); -1218final ArrayListRegionInfo regionsToAssign = new ArrayListRegionInfo(); -1219long st, et; -1220 -1221st = System.currentTimeMillis(); -1222for (RegionStateNode regionNode: regionStates.getRegionNodes()) { -1223 if (regionNode.getState() == State.OPEN) { -1224final ServerName serverName = regionNode.getRegionLocation(); -1225if (!master.getServerManager().isServerOnline(serverName)) { -1226 offlineServersWithOnlineRegions.add(serverName); -1227} -1228 } else if (regionNode.getState() == State.OFFLINE) { -1229if (isTableEnabled(regionNode.getTable())) { -1230 regionsToAssign.add(regionNode.getRegionInfo()); -1231} -1232 } -1233} -1234et = System.currentTimeMillis(); -1235LOG.info("[STEP-1] " + StringUtils.humanTimeDiff(et - st)); -1236 -1237// kill servers with online regions -1238st = System.currentTimeMillis(); -1239for (ServerName serverName: offlineServersWithOnlineRegions) { -1240 if (!master.getServerManager().isServerOnline(serverName)) { -1241LOG.info("KILL RS hosting regions but not online " + serverName + -1242 " (master=" + master.getServerName() + ")"); -1243killRegionServer(serverName); -1244 } -1245} -1246et = System.currentTimeMillis(); -1247LOG.info("[STEP-2] " + StringUtils.humanTimeDiff(et - st)); -1248 -1249setFailoverCleanupDone(true); -1250 -1251// assign offline regions -1252
[06/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode,
[20/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html index 8e0a6a2..08916af 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.ServerReportEvent.html @@ -75,7 +75,7 @@ 067 }; 068 069 protected static final State[] STATES_EXPECTED_ON_CLOSE = new State[] { -070State.SPLITTING, State.SPLIT, // ServerCrashProcedure +070State.SPLITTING, State.SPLIT, State.MERGING, // ServerCrashProcedure 071State.OPEN, // enabled/open 072State.CLOSING // already in-progress (retrying) 073 }; @@ -550,446 +550,454 @@ 542return !getTableRegionStates(tableName).isEmpty(); 543 } 544 -545 public ListRegionInfo getRegionsOfTable(final TableName table) { -546return getRegionsOfTable(table, false); -547 } -548 -549 ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { -550final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); -551final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size()); -552for (RegionStateNode node: nodes) { -553 if (include(node, offline)) hris.add(node.getRegionInfo()); -554} -555return hris; -556 } -557 -558 /** -559 * Utility. Whether to include region in list of regions. Default is to -560 * weed out split and offline regions. -561 * @return True if we should include the codenode/code (do not include -562 * if split or offline unless codeoffline/code is set to true. -563 */ -564 boolean include(final RegionStateNode node, final boolean offline) { -565if (LOG.isTraceEnabled()) { -566 LOG.trace("WORKING ON " + node + " " + node.getRegionInfo()); -567} -568if (node.isInState(State.SPLIT)) return false; -569if (node.isInState(State.OFFLINE) !offline) return false; -570final RegionInfo hri = node.getRegionInfo(); -571return (!hri.isOffline() !hri.isSplit()) || -572((hri.isOffline() || hri.isSplit()) offline); -573 } -574 -575 /** -576 * Returns the set of regions hosted by the specified server -577 * @param serverName the server we are interested in -578 * @return set of RegionInfo hosted by the specified server -579 */ -580 public ListRegionInfo getServerRegionInfoSet(final ServerName serverName) { -581final ServerStateNode serverInfo = getServerNode(serverName); -582if (serverInfo == null) return Collections.emptyList(); -583 -584synchronized (serverInfo) { -585 return serverInfo.getRegionInfoList(); -586} -587 } -588 -589 // -590 // TODO: split helpers -591 // -592 public void logSplit(final ServerName serverName) { -593final ServerStateNode serverNode = getOrCreateServer(serverName); -594synchronized (serverNode) { -595 serverNode.setState(ServerState.SPLITTING); -596 /* THIS HAS TO BE WRONG. THIS IS SPLITTING OF REGION, NOT SPLITTING WALs. -597 for (RegionStateNode regionNode: serverNode.getRegions()) { -598synchronized (regionNode) { -599 // TODO: Abort procedure if present -600 regionNode.setState(State.SPLITTING); -601} -602 }*/ -603} -604 } -605 -606 public void logSplit(final RegionInfo regionInfo) { -607final RegionStateNode regionNode = getRegionNode(regionInfo); -608synchronized (regionNode) { -609 regionNode.setState(State.SPLIT); -610} -611 } -612 -613 @VisibleForTesting -614 public void updateRegionState(final RegionInfo regionInfo, final State state) { -615final RegionStateNode regionNode = getOrCreateRegionNode(regionInfo); +545 /** +546 * @return Return online regions of table; does not include OFFLINE or SPLITTING regions. +547 */ +548 public ListRegionInfo getRegionsOfTable(final TableName table) { +549return getRegionsOfTable(table, false); +550 } +551 +552 /** +553 * @return Return the regions of the table; does not include OFFLINE unless you set +554 * codeoffline/code to true. Does not include regions that are in the +555 * {@link State#SPLIT} state. +556 */ +557 public ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { +558final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); +559final ArrayListRegionInfo hris = new
[03/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode,
[01/51] [partial] hbase-site git commit: Published site at .
Repository: hbase-site Updated Branches: refs/heads/asf-site d0ae29f24 -> a108018f1 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is
[13/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode,
[02/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode, FSWALEntry#stampRegionSequenceId won't stamp sequence
[26/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html index 845f5e6..84e9e52 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStateStore.RegionStateVisitor.html @@ -98,241 +98,248 @@ 090 @Override 091 public boolean visit(final Result r) throws IOException { 092if (r != null !r.isEmpty()) { -093 long st = System.currentTimeMillis(); -094 visitMetaEntry(visitor, r); -095 long et = System.currentTimeMillis(); -096 LOG.info("[T] LOAD META PERF " + StringUtils.humanTimeDiff(et - st)); -097} else if (isDebugEnabled) { -098 LOG.debug("NULL result from meta - ignoring but this is strange."); -099} -100return true; -101 } -102}); -103 } -104 -105 private void visitMetaEntry(final RegionStateVisitor visitor, final Result result) -106 throws IOException { -107final RegionLocations rl = MetaTableAccessor.getRegionLocations(result); -108if (rl == null) return; +093 long st = 0; +094 if (LOG.isTraceEnabled()) { +095st = System.currentTimeMillis(); +096 } +097 visitMetaEntry(visitor, r); +098 if (LOG.isTraceEnabled()) { +099long et = System.currentTimeMillis(); +100LOG.trace("[T] LOAD META PERF " + StringUtils.humanTimeDiff(et - st)); +101 } +102} else if (isDebugEnabled) { +103 LOG.debug("NULL result from meta - ignoring but this is strange."); +104} +105return true; +106 } +107}); +108 } 109 -110final HRegionLocation[] locations = rl.getRegionLocations(); -111if (locations == null) return; -112 -113for (int i = 0; i locations.length; ++i) { -114 final HRegionLocation hrl = locations[i]; -115 if (hrl == null) continue; -116 -117 final RegionInfo regionInfo = hrl.getRegionInfo(); -118 if (regionInfo == null) continue; -119 -120 final int replicaId = regionInfo.getReplicaId(); -121 final State state = getRegionState(result, replicaId); -122 -123 final ServerName lastHost = hrl.getServerName(); -124 final ServerName regionLocation = getRegionServer(result, replicaId); -125 final long openSeqNum = -1; -126 -127 // TODO: move under trace, now is visible for debugging -128 LOG.info(String.format("Load hbase:meta entry region=%s regionState=%s lastHost=%s regionLocation=%s", -129regionInfo, state, lastHost, regionLocation)); -130 -131 visitor.visitRegionState(regionInfo, state, regionLocation, lastHost, openSeqNum); -132} -133 } -134 -135 public void updateRegionLocation(final RegionInfo regionInfo, final State state, -136 final ServerName regionLocation, final ServerName lastHost, final long openSeqNum, -137 final long pid) -138 throws IOException { -139if (regionInfo.isMetaRegion()) { -140 updateMetaLocation(regionInfo, regionLocation); -141} else { -142 updateUserRegionLocation(regionInfo, state, regionLocation, lastHost, openSeqNum, pid); -143} -144 } -145 -146 public void updateRegionState(final long openSeqNum, final long pid, -147 final RegionState newState, final RegionState oldState) throws IOException { -148 updateRegionLocation(newState.getRegion(), newState.getState(), newState.getServerName(), -149oldState != null ? oldState.getServerName() : null, openSeqNum, pid); -150 } -151 -152 protected void updateMetaLocation(final RegionInfo regionInfo, final ServerName serverName) -153 throws IOException { -154try { -155 MetaTableLocator.setMetaLocation(master.getZooKeeper(), serverName, -156regionInfo.getReplicaId(), State.OPEN); -157} catch (KeeperException e) { -158 throw new IOException(e); -159} -160 } -161 -162 protected void updateUserRegionLocation(final RegionInfo regionInfo, final State state, -163 final ServerName regionLocation, final ServerName lastHost, final long openSeqNum, -164 final long pid) -165 throws IOException { -166final int replicaId = regionInfo.getReplicaId(); -167final Put put = new Put(MetaTableAccessor.getMetaKeyForRegion(regionInfo)); -168MetaTableAccessor.addRegionInfo(put, regionInfo); -169final StringBuilder info = new StringBuilder("pid=" + pid + " updating hbase:meta row="); -170 info.append(regionInfo.getRegionNameAsString()).append(",
[04/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode, FSWALEntry#stampRegionSequenceId won't
[08/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode, FSWALEntry#stampRegionSequenceId won't stamp
[11/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode, FSWALEntry#stampRegionSequenceId won't stamp
[14/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html index ffb1464..da77107 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.html @@ -6,7 +6,7 @@ -001/** +001/* 002 * Licensed to the Apache Software Foundation (ASF) under one 003 * or more contributor license agreements. See the NOTICE file 004 * distributed with this work for additional information @@ -128,9 +128,9 @@ 120// Get the replica count 121int regionReplicaCount = hTableDescriptor.getRegionReplication(); 122 -123// Get the regions for the table from the memory +123// Get the regions for the table from memory; get both online and offline regions ('true'). 124ListRegionInfo regionsOfTable = -125 env.getAssignmentManager().getRegionStates().getRegionsOfTable(tableName); +125 env.getAssignmentManager().getRegionStates().getRegionsOfTable(tableName, true); 126 127if (regionReplicaCount 1) { 128 int currentMaxReplica = 0;
[10/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode, FSWALEntry#stampRegionSequenceId
[12/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode, FSWALEntry#stampRegionSequenceId won't stamp sequence
[21/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html index 8e0a6a2..08916af 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html @@ -75,7 +75,7 @@ 067 }; 068 069 protected static final State[] STATES_EXPECTED_ON_CLOSE = new State[] { -070State.SPLITTING, State.SPLIT, // ServerCrashProcedure +070State.SPLITTING, State.SPLIT, State.MERGING, // ServerCrashProcedure 071State.OPEN, // enabled/open 072State.CLOSING // already in-progress (retrying) 073 }; @@ -550,446 +550,454 @@ 542return !getTableRegionStates(tableName).isEmpty(); 543 } 544 -545 public ListRegionInfo getRegionsOfTable(final TableName table) { -546return getRegionsOfTable(table, false); -547 } -548 -549 ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { -550final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); -551final ArrayListRegionInfo hris = new ArrayListRegionInfo(nodes.size()); -552for (RegionStateNode node: nodes) { -553 if (include(node, offline)) hris.add(node.getRegionInfo()); -554} -555return hris; -556 } -557 -558 /** -559 * Utility. Whether to include region in list of regions. Default is to -560 * weed out split and offline regions. -561 * @return True if we should include the codenode/code (do not include -562 * if split or offline unless codeoffline/code is set to true. -563 */ -564 boolean include(final RegionStateNode node, final boolean offline) { -565if (LOG.isTraceEnabled()) { -566 LOG.trace("WORKING ON " + node + " " + node.getRegionInfo()); -567} -568if (node.isInState(State.SPLIT)) return false; -569if (node.isInState(State.OFFLINE) !offline) return false; -570final RegionInfo hri = node.getRegionInfo(); -571return (!hri.isOffline() !hri.isSplit()) || -572((hri.isOffline() || hri.isSplit()) offline); -573 } -574 -575 /** -576 * Returns the set of regions hosted by the specified server -577 * @param serverName the server we are interested in -578 * @return set of RegionInfo hosted by the specified server -579 */ -580 public ListRegionInfo getServerRegionInfoSet(final ServerName serverName) { -581final ServerStateNode serverInfo = getServerNode(serverName); -582if (serverInfo == null) return Collections.emptyList(); -583 -584synchronized (serverInfo) { -585 return serverInfo.getRegionInfoList(); -586} -587 } -588 -589 // -590 // TODO: split helpers -591 // -592 public void logSplit(final ServerName serverName) { -593final ServerStateNode serverNode = getOrCreateServer(serverName); -594synchronized (serverNode) { -595 serverNode.setState(ServerState.SPLITTING); -596 /* THIS HAS TO BE WRONG. THIS IS SPLITTING OF REGION, NOT SPLITTING WALs. -597 for (RegionStateNode regionNode: serverNode.getRegions()) { -598synchronized (regionNode) { -599 // TODO: Abort procedure if present -600 regionNode.setState(State.SPLITTING); -601} -602 }*/ -603} -604 } -605 -606 public void logSplit(final RegionInfo regionInfo) { -607final RegionStateNode regionNode = getRegionNode(regionInfo); -608synchronized (regionNode) { -609 regionNode.setState(State.SPLIT); -610} -611 } -612 -613 @VisibleForTesting -614 public void updateRegionState(final RegionInfo regionInfo, final State state) { -615final RegionStateNode regionNode = getOrCreateRegionNode(regionInfo); +545 /** +546 * @return Return online regions of table; does not include OFFLINE or SPLITTING regions. +547 */ +548 public ListRegionInfo getRegionsOfTable(final TableName table) { +549return getRegionsOfTable(table, false); +550 } +551 +552 /** +553 * @return Return the regions of the table; does not include OFFLINE unless you set +554 * codeoffline/code to true. Does not include regions that are in the +555 * {@link State#SPLIT} state. +556 */ +557 public ListRegionInfo getRegionsOfTable(final TableName table, final boolean offline) { +558final ArrayListRegionStateNode nodes = getTableRegionStateNodes(table); +559
[09/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a108018f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html index 5c5e0eb..7b41db2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html @@ -2960,5416 +2960,5447 @@ 2952protected final ObservedExceptionsInBatch observedExceptions; 2953//Durability of the batch (highest durability of all operations) 2954protected Durability durability; -2955 -2956public BatchOperation(final HRegion region, T[] operations) { -2957 this.operations = operations; -2958 this.retCodeDetails = new OperationStatus[operations.length]; -2959 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -2960 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -2961 familyCellMaps = new Map[operations.length]; -2962 -2963 this.region = region; -2964 observedExceptions = new ObservedExceptionsInBatch(); -2965 durability = Durability.USE_DEFAULT; -2966} -2967 -2968/** -2969 * Visitor interface for batch operations -2970 */ -2971@FunctionalInterface -2972public interface Visitor { -2973 /** -2974 * @param index operation index -2975 * @return If true continue visiting remaining entries, break otherwise -2976 */ -2977 boolean visit(int index) throws IOException; -2978} -2979 -2980/** -2981 * Helper method for visiting pending/ all batch operations -2982 */ -2983public void visitBatchOperations(boolean pendingOnly, int lastIndexExclusive, Visitor visitor) -2984throws IOException { -2985 assert lastIndexExclusive = this.size(); -2986 for (int i = nextIndexToProcess; i lastIndexExclusive; i++) { -2987if (!pendingOnly || isOperationPending(i)) { -2988 if (!visitor.visit(i)) { -2989break; -2990 } -2991} -2992 } -2993} -2994 -2995public abstract Mutation getMutation(int index); -2996public abstract long getNonceGroup(int index); -2997public abstract long getNonce(int index); -2998/** This method is potentially expensive and useful mostly for non-replay CP path. */ -2999public abstract Mutation[] getMutationsForCoprocs(); -3000public abstract boolean isInReplay(); -3001public abstract long getOrigLogSeqNum(); -3002public abstract void startRegionOperation() throws IOException; -3003public abstract void closeRegionOperation() throws IOException; -3004 -3005/** -3006 * Validates each mutation and prepares a batch for write. If necessary (non-replay case), runs -3007 * CP prePut()/ preDelete() hooks for all mutations in a batch. This is intended to operate on -3008 * entire batch and will be called from outside of class to check and prepare batch. This can -3009 * be implemented by calling helper method {@link #checkAndPrepareMutation(int, long)} in a -3010 * 'for' loop over mutations. -3011 */ -3012public abstract void checkAndPrepare() throws IOException; -3013 -3014/** -3015 * Implement any Put request specific check and prepare logic here. Please refer to -3016 * {@link #checkAndPrepareMutation(Mutation, long)} for how its used. -3017 */ -3018protected abstract void checkAndPreparePut(final Put p) throws IOException; -3019 -3020/** -3021 * If necessary, calls preBatchMutate() CP hook for a mini-batch and updates metrics, cell -3022 * count, tags and timestamp for all cells of all operations in a mini-batch. -3023 */ -3024public abstract void prepareMiniBatchOperations(MiniBatchOperationInProgressMutation -3025miniBatchOp, long timestamp, final ListRowLock acquiredRowLocks) throws IOException; -3026 -3027/** -3028 * Write mini-batch operations to MemStore -3029 */ -3030public abstract WriteEntry writeMiniBatchOperationsToMemStore( -3031final MiniBatchOperationInProgressMutation miniBatchOp, final WriteEntry writeEntry) -3032throws IOException; -3033 -3034protected void writeMiniBatchOperationsToMemStore( -3035final MiniBatchOperationInProgressMutation miniBatchOp, final long writeNumber) -3036throws IOException { -3037 MemStoreSizing memStoreAccounting = new MemStoreSizing(); -3038 visitBatchOperations(true, miniBatchOp.getLastIndexExclusive(), (int index) - { -3039// We need to update the sequence id for following reasons. -3040// 1) If the op is in replay mode, FSWALEntry#stampRegionSequenceId won't stamp sequence id. -3041
hbase git commit: HBASE-16459 Remove unused hbase shell --format option
Repository: hbase Updated Branches: refs/heads/master 35a1e4334 -> 31234eb86 HBASE-16459 Remove unused hbase shell --format option Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/31234eb8 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/31234eb8 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/31234eb8 Branch: refs/heads/master Commit: 31234eb862fdd7ee4917a2f74001182565ffbfd9 Parents: 35a1e43 Author: Apekshit SharmaAuthored: Thu Nov 9 15:03:13 2017 -0800 Committer: Apekshit Sharma Committed: Fri Nov 10 00:08:02 2017 -0800 -- bin/hirb.rb | 17 + 1 file changed, 1 insertion(+), 16 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/31234eb8/bin/hirb.rb -- diff --git a/bin/hirb.rb b/bin/hirb.rb index c06c83b..35895f8 100644 --- a/bin/hirb.rb +++ b/bin/hirb.rb @@ -54,10 +54,6 @@ $LOAD_PATH.unshift Pathname.new(sources) cmdline_help = <