[04/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftTable.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftTable.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftTable.html new file mode 100644 index 000..94888eb --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift2/client/ThriftTable.html @@ -0,0 +1,564 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * +003 * Licensed to the Apache Software Foundation (ASF) under one +004 * or more contributor license agreements. See the NOTICE file +005 * distributed with this work for additional information +006 * regarding copyright ownership. The ASF licenses this file +007 * to you under the Apache License, Version 2.0 (the +008 * "License"); you may not use this file except in compliance +009 * with the License. You may obtain a copy of the License at +010 * +011 * http://www.apache.org/licenses/LICENSE-2.0 +012 * +013 * Unless required by applicable law or agreed to in writing, software +014 * distributed under the License is distributed on an "AS IS" BASIS, +015 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +016 * See the License for the specific language governing permissions and +017 * limitations under the License. +018 */ +019package org.apache.hadoop.hbase.thrift2.client; +020 +021import static org.apache.hadoop.hbase.thrift.Constants.HBASE_THRIFT_CLIENT_SCANNER_CACHING; +022import static org.apache.hadoop.hbase.thrift.Constants.HBASE_THRIFT_CLIENT_SCANNER_CACHING_DEFAULT; +023 +024import java.io.IOException; +025import java.nio.ByteBuffer; +026import java.util.ArrayDeque; +027import java.util.ArrayList; +028import java.util.Arrays; +029import java.util.List; +030import java.util.Queue; +031import java.util.concurrent.TimeUnit; +032 +033import org.apache.commons.lang3.NotImplementedException; +034import org.apache.hadoop.conf.Configuration; +035import org.apache.hadoop.hbase.CompareOperator; +036import org.apache.hadoop.hbase.HConstants; +037import org.apache.hadoop.hbase.TableName; +038import org.apache.hadoop.hbase.client.Append; +039import org.apache.hadoop.hbase.client.Delete; +040import org.apache.hadoop.hbase.client.Get; +041import org.apache.hadoop.hbase.client.Increment; +042import org.apache.hadoop.hbase.client.Put; +043import org.apache.hadoop.hbase.client.Result; +044import org.apache.hadoop.hbase.client.ResultScanner; +045import org.apache.hadoop.hbase.client.Row; +046import org.apache.hadoop.hbase.client.RowMutations; +047import org.apache.hadoop.hbase.client.Scan; +048import org.apache.hadoop.hbase.client.Table; +049import org.apache.hadoop.hbase.client.TableDescriptor; +050import org.apache.hadoop.hbase.client.coprocessor.Batch; +051import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +052import org.apache.hadoop.hbase.io.TimeRange; +053import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; +054import org.apache.hadoop.hbase.thrift2.ThriftUtilities; +055import org.apache.hadoop.hbase.thrift2.generated.TAppend; +056import org.apache.hadoop.hbase.thrift2.generated.TDelete; +057import org.apache.hadoop.hbase.thrift2.generated.TGet; +058import org.apache.hadoop.hbase.thrift2.generated.THBaseService; +059import org.apache.hadoop.hbase.thrift2.generated.TIncrement; +060import org.apache.hadoop.hbase.thrift2.generated.TPut; +061import org.apache.hadoop.hbase.thrift2.generated.TResult; +062import org.apache.hadoop.hbase.thrift2.generated.TRowMutations; +063import org.apache.hadoop.hbase.thrift2.generated.TScan; +064import org.apache.hadoop.hbase.thrift2.generated.TTableDescriptor; +065import org.apache.hadoop.hbase.util.Bytes; +066import org.apache.thrift.TException; +067import org.apache.thrift.transport.TTransport; +068import org.apache.yetus.audience.InterfaceAudience; +069 +070import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +071import org.apache.hbase.thirdparty.com.google.common.primitives.Booleans; +072 +073@InterfaceAudience.Private +074public class ThriftTable implements Table { +075 +076 private TableName tableName; +077 private Configuration conf; +078 private TTransport tTransport; +079 private THBaseService.Client client; +080 private ByteBuffer tableNameInBytes; +081 private int operationTimeout; +082 +083 private final int scannerCaching; +084 +085 public ThriftTable(TableName tableName, THBaseService.Client client, TTransport tTransport, +086 Configuration conf) { +087this.tableName = tableName; +088this.tableNameInBytes = ByteBuffer.wrap(tableName.toBytes()); +089this.conf = conf; +090this.tTransport = tTransport; +091this.client = client; +092this.scannerCaching = conf.getInt(HBASE_THRIFT_CLIENT_SCANNER_CACHING, +093
[04/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.html b/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.html index c69f626..7ac136f 100644 --- a/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.html +++ b/devapidocs/org/apache/hadoop/hbase/rsgroup/RSGroupAdminEndpoint.html @@ -347,7 +347,7 @@ implements MasterObserver -postAbortProcedure, postAddReplicationPeer, postAddRSGroup, postAssign, post Balance, postBalanceRSGroup, postBalanceSwitch, postCloneSnapshot, postCompletedCreateTableAction, postCompletedDeleteTableAction, postCompletedDisableTableAction, postCompletedEnableTableAction, postCompletedMergeRegionsAction, postCompletedModifyTableAction, postCompletedModifyTableAction, postCompletedSplitRegionAction, postCompletedTruncateTableAction, postCreateNamespace, postDecommissionRegionServers, postDeleteNamespace, postDeleteSnapshot, postDisableReplicationPeer, postDisableTable, postEnableReplicationPeer, postEnableTable, postGetClusterMetrics, postGetLocks, postGetNamespaceDescriptor, postGetProcedures, postGetReplicationPeerConfig, postGetRSGroupInfo, postGetRSGroupInfoOfServer, postGetRSGroupInfoOfTable, postGetTableDescriptors, postGetTableNames, postListDecommissionedRegionServers, postListNamespaceDescriptors, postListReplicationPeers, postListRSGroups, postListSnapshot, postLockHeartbeat, postMergeRegions, postMergeRegionsCommitAction, postModifyNamespace, postModifyNamespace, postModifyTable, postModifyTable, postMove, postMoveServers, postMoveServersAndTables, postMoveTables, postRecommissionRegionServer, postRegionOffline, postRemoveReplicationPeer, postRemoveRSGroup, postRemoveServers, postRequestLock, postRestoreSnapshot, postRollBackMergeRegionsAction, postRollBackSplitRegionAction, postSetNamespaceQuota, postSetSplitOrMergeEnabled, postSetTableQuota, postSetUserQuota, postSetUserQuota, postSetUserQuota, postSnapshot, postStartMaster, postTableFlush, postTransitReplicationPeerSyncReplicationState, postTruncateTable, postUnassign, postUpdateReplicationPeerConfig, preAbortProcedure, preAddReplicationPeer, preAddRSGroup, preAssign, preBalance, preBalanceRSGroup, preBalanceSwitch, preClearDeadServers, preCreateTable, preCreateTableRegionsInfos, preDecommissionRegionServers, preDeleteNamespace, preDeleteSnapshot, preDeleteTable, preDeleteTableAction, preDisableReplicationPeer, preDisableTable, preDisableTableAction, preEnableReplicationPeer, preEnableTable, preEnableTableAction, preGetClusterMetrics, preGetLocks, preGetNamespaceDescriptor, preGetProcedures, preGetReplicationPeerConfig, preGetRSGroupInfo, preGetRSGroupInfoOfServer, preGetRSGroupInfoOfTable, preGetTableDescriptors, preGetTableNames, preListDecommissionedRegionServers, preListNamespaceDescriptors, preListReplicationPeers, preListRSGroups, preListSnapshot, preLockHeartbeat, preMasterInitialization, preMergeRegions, preMergeRegionsAction, preMergeRegionsCommitAction, preModifyNamespace, preModifyTable, preModifyTable, preModifyTableAction, preModifyTableAction, preMove, preMoveServers, preMoveServersAndTables, preMoveTables, preRecommissionRegionServer, preRegionOffline, preRemoveReplicationPeer, preRemoveRSGroup, preRemoveServers, preRequestLock, preRestoreSnapshot, preSetNamespaceQuota, preSetSplitOrMergeEnabled, preSetTableQuot a, preSetUserQuota, preSetUserQuota, preSetUserQuota, preShutdown, preSnapshot, preSplitRegion, preSplitRegionAction, preSplitRegionAfterMETAAction, preSplitRegionBeforeMETAAction, preStopMaster, preTableFlush, preTransitReplicationPeerSyncReplicationState, preTruncateTable, preTruncateTableAction, preUnassign, preUpdateReplicationPeerConfig +postAbortProcedure, postAddReplicationPeer, postAddRSGroup, postAssign, post Balance, postBalanceRSGroup, postBalanceSwitch, postCloneSnapshot, postCompletedCreateTableAction, postCompletedDeleteTableAction, postCompletedDisableTableAction, postCompletedEnableTableAction, postCompletedMergeRegionsAction, postCompletedModifyTableAction, postCompletedModifyTableAction, postCompletedSplitRegionAction, postCompletedTruncateTableAction, postCreateNamespace, postDecommissionRegionServers, postDeleteNamespace, postDeleteSnapshot, postDisableReplicationPeer, postDisableTable, postEnableReplicationPeer, postEnableTable, postGetClusterMetrics, postGetLocks, postGetNamespaceDescriptor, postGetProcedures, postGetReplicationPeerConfig, postGetRSGroupInfo, postGetRSGroupInfoOfServer, postGetRSGroupInfoOfTable, postGetTableDescriptors, postGetTableNames, postIsRpcThrottleEnabled, postListDecommissionedRegionServers, postListNamespaceDescriptors,
[04/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html index 736388b..197b99d 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html @@ -26,3624 +26,3599 @@ 018package org.apache.hadoop.hbase.client; 019 020import static org.apache.hadoop.hbase.TableName.META_TABLE_NAME; -021 -022import com.google.protobuf.Message; -023import com.google.protobuf.RpcChannel; -024import java.io.IOException; -025import java.util.ArrayList; -026import java.util.Arrays; -027import java.util.Collections; -028import java.util.EnumSet; -029import java.util.HashMap; -030import java.util.List; -031import java.util.Map; -032import java.util.Optional; -033import java.util.Set; -034import java.util.concurrent.CompletableFuture; -035import java.util.concurrent.ConcurrentHashMap; -036import java.util.concurrent.TimeUnit; -037import java.util.concurrent.atomic.AtomicReference; -038import java.util.function.BiConsumer; -039import java.util.function.Function; -040import java.util.function.Supplier; -041import java.util.regex.Pattern; -042import java.util.stream.Collectors; -043import java.util.stream.Stream; -044import org.apache.commons.io.IOUtils; -045import org.apache.hadoop.conf.Configuration; -046import org.apache.hadoop.hbase.AsyncMetaTableAccessor; -047import org.apache.hadoop.hbase.CacheEvictionStats; -048import org.apache.hadoop.hbase.CacheEvictionStatsAggregator; -049import org.apache.hadoop.hbase.ClusterMetrics; -050import org.apache.hadoop.hbase.ClusterMetrics.Option; -051import org.apache.hadoop.hbase.ClusterMetricsBuilder; -052import org.apache.hadoop.hbase.HConstants; -053import org.apache.hadoop.hbase.HRegionLocation; -054import org.apache.hadoop.hbase.MetaTableAccessor; -055import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -056import org.apache.hadoop.hbase.NamespaceDescriptor; -057import org.apache.hadoop.hbase.RegionLocations; -058import org.apache.hadoop.hbase.RegionMetrics; -059import org.apache.hadoop.hbase.RegionMetricsBuilder; -060import org.apache.hadoop.hbase.ServerName; -061import org.apache.hadoop.hbase.TableExistsException; -062import org.apache.hadoop.hbase.TableName; -063import org.apache.hadoop.hbase.TableNotDisabledException; -064import org.apache.hadoop.hbase.TableNotEnabledException; -065import org.apache.hadoop.hbase.TableNotFoundException; -066import org.apache.hadoop.hbase.UnknownRegionException; -067import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder; -068import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder; -069import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.ServerRequestCallerBuilder; -070import org.apache.hadoop.hbase.client.Scan.ReadType; -071import org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil; -072import org.apache.hadoop.hbase.client.replication.TableCFs; -073import org.apache.hadoop.hbase.client.security.SecurityCapability; -074import org.apache.hadoop.hbase.exceptions.DeserializationException; -075import org.apache.hadoop.hbase.ipc.HBaseRpcController; -076import org.apache.hadoop.hbase.quotas.QuotaFilter; -077import org.apache.hadoop.hbase.quotas.QuotaSettings; -078import org.apache.hadoop.hbase.quotas.QuotaTableUtil; -079import org.apache.hadoop.hbase.replication.ReplicationException; -080import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -081import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -082import org.apache.hadoop.hbase.replication.SyncReplicationState; -083import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; -084import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException; -085import org.apache.hadoop.hbase.snapshot.SnapshotCreationException; -086import org.apache.hadoop.hbase.util.Bytes; -087import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -088import org.apache.hadoop.hbase.util.ForeignExceptionUtil; -089import org.apache.yetus.audience.InterfaceAudience; -090import org.slf4j.Logger; -091import org.slf4j.LoggerFactory; -092 -093import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -094import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; -095import org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback; -096import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; -097import
[04/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/testdevapidocs/org/apache/hadoop/hbase/thrift2/TestThrift2ServerCmdLine.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/thrift2/TestThrift2ServerCmdLine.html b/testdevapidocs/org/apache/hadoop/hbase/thrift2/TestThrift2ServerCmdLine.html new file mode 100644 index 000..54f2a06 --- /dev/null +++ b/testdevapidocs/org/apache/hadoop/hbase/thrift2/TestThrift2ServerCmdLine.html @@ -0,0 +1,374 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +TestThrift2ServerCmdLine (Apache HBase 3.0.0-SNAPSHOT Test API) + + + + + +var methods = {"i0":10,"i1":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.thrift2 +Class TestThrift2ServerCmdLine + + + +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine + + +org.apache.hadoop.hbase.thrift2.TestThrift2ServerCmdLine + + + + + + + + + + +public class TestThrift2ServerCmdLine +extends TestThriftServerCmdLine + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +static HBaseClassTestRule +CLASS_RULE + + +private static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +TABLENAME + + + + + + +Fields inherited from classorg.apache.hadoop.hbase.thrift.TestThriftServerCmdLine +implType, port, specifyBindIP, specifyCompact, specifyFramed, tableCreated, TEST_UTIL + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +TestThrift2ServerCmdLine(org.apache.hadoop.hbase.thrift.ImplTypeimplType, +booleanspecifyFramed, +booleanspecifyBindIP, +booleanspecifyCompact) + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +protected org.apache.hadoop.hbase.thrift2.ThriftServer +createThriftServer() + + +protected void +talkToThriftServer() + + + + + + +Methods inherited from classorg.apache.hadoop.hbase.thrift.TestThriftServerCmdLine +getParameters, setUpBeforeClass, tearDownAfterClass, testRunThriftServer + + + + + +Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/ja va/lang/Object.html?is-external=true#notifyAll--" title="class or interface in java.lang">notifyAll, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--; title="class or interface in java.lang">toString, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--; title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-; title="class or interface in java.lang">wait,
[04/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html index 8ba9cf5..fbf978f 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStateStore.html @@ -712,6 +712,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html index dff40e3..e5f288e 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html @@ -398,6 +398,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html index 1b64578..4401a5a 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionStateStampComparator.html @@ -291,6 +291,6 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/Comparat -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html index b1366ce..8b8259c 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.html @@ -1319,6 +1319,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html index 7998852..8dfeade 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html @@ -970,6 +970,6 @@ implements Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/master/assignment/ServerState.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/ServerState.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/ServerState.html index 6837096..d9b370b 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/ServerState.html +++
[04/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html index 0f5a095..50bf692 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html @@ -78,8712 +78,8714 @@ 070import java.util.concurrent.locks.ReadWriteLock; 071import java.util.concurrent.locks.ReentrantReadWriteLock; 072import java.util.function.Function; -073import org.apache.hadoop.conf.Configuration; -074import org.apache.hadoop.fs.FileStatus; -075import org.apache.hadoop.fs.FileSystem; -076import org.apache.hadoop.fs.LocatedFileStatus; -077import org.apache.hadoop.fs.Path; -078import org.apache.hadoop.hbase.Cell; -079import org.apache.hadoop.hbase.CellBuilderType; -080import org.apache.hadoop.hbase.CellComparator; -081import org.apache.hadoop.hbase.CellComparatorImpl; -082import org.apache.hadoop.hbase.CellScanner; -083import org.apache.hadoop.hbase.CellUtil; -084import org.apache.hadoop.hbase.CompareOperator; -085import org.apache.hadoop.hbase.CompoundConfiguration; -086import org.apache.hadoop.hbase.DoNotRetryIOException; -087import org.apache.hadoop.hbase.DroppedSnapshotException; -088import org.apache.hadoop.hbase.ExtendedCellBuilderFactory; -089import org.apache.hadoop.hbase.HConstants; -090import org.apache.hadoop.hbase.HConstants.OperationStatusCode; -091import org.apache.hadoop.hbase.HDFSBlocksDistribution; -092import org.apache.hadoop.hbase.KeyValue; -093import org.apache.hadoop.hbase.KeyValueUtil; -094import org.apache.hadoop.hbase.NamespaceDescriptor; -095import org.apache.hadoop.hbase.NotServingRegionException; -096import org.apache.hadoop.hbase.PrivateCellUtil; -097import org.apache.hadoop.hbase.RegionTooBusyException; -098import org.apache.hadoop.hbase.Tag; -099import org.apache.hadoop.hbase.TagUtil; -100import org.apache.hadoop.hbase.UnknownScannerException; -101import org.apache.hadoop.hbase.client.Append; -102import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -103import org.apache.hadoop.hbase.client.CompactionState; -104import org.apache.hadoop.hbase.client.Delete; -105import org.apache.hadoop.hbase.client.Durability; -106import org.apache.hadoop.hbase.client.Get; -107import org.apache.hadoop.hbase.client.Increment; -108import org.apache.hadoop.hbase.client.IsolationLevel; -109import org.apache.hadoop.hbase.client.Mutation; -110import org.apache.hadoop.hbase.client.PackagePrivateFieldAccessor; -111import org.apache.hadoop.hbase.client.Put; -112import org.apache.hadoop.hbase.client.RegionInfo; -113import org.apache.hadoop.hbase.client.RegionInfoBuilder; -114import org.apache.hadoop.hbase.client.RegionReplicaUtil; -115import org.apache.hadoop.hbase.client.Result; -116import org.apache.hadoop.hbase.client.RowMutations; -117import org.apache.hadoop.hbase.client.Scan; -118import org.apache.hadoop.hbase.client.TableDescriptor; -119import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -120import org.apache.hadoop.hbase.conf.ConfigurationManager; -121import org.apache.hadoop.hbase.conf.PropagatingConfigurationObserver; -122import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -123import org.apache.hadoop.hbase.coprocessor.RegionObserver.MutationType; -124import org.apache.hadoop.hbase.errorhandling.ForeignExceptionSnare; -125import org.apache.hadoop.hbase.exceptions.FailedSanityCheckException; -126import org.apache.hadoop.hbase.exceptions.TimeoutIOException; -127import org.apache.hadoop.hbase.exceptions.UnknownProtocolException; -128import org.apache.hadoop.hbase.filter.ByteArrayComparable; -129import org.apache.hadoop.hbase.filter.FilterWrapper; -130import org.apache.hadoop.hbase.filter.IncompatibleFilterException; -131import org.apache.hadoop.hbase.io.HFileLink; -132import org.apache.hadoop.hbase.io.HeapSize; -133import org.apache.hadoop.hbase.io.TimeRange; -134import org.apache.hadoop.hbase.io.hfile.BlockCache; -135import org.apache.hadoop.hbase.io.hfile.HFile; -136import org.apache.hadoop.hbase.ipc.CallerDisconnectedException; -137import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -138import org.apache.hadoop.hbase.ipc.RpcCall; -139import org.apache.hadoop.hbase.ipc.RpcServer; -140import org.apache.hadoop.hbase.mob.MobFileCache; -141import org.apache.hadoop.hbase.monitoring.MonitoredTask; -142import org.apache.hadoop.hbase.monitoring.TaskMonitor; -143import org.apache.hadoop.hbase.quotas.RegionServerSpaceQuotaManager; -144import org.apache.hadoop.hbase.regionserver.MultiVersionConcurrencyControl.WriteEntry; -145import
[04/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.html index 79cb21b..d8d391b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.html @@ -378,1508 +378,1510 @@ 370 371 @Override 372 public void returnBlock(HFileBlock block) { -373BlockCache blockCache = this.cacheConf.getBlockCache(); -374if (blockCache != null block != null) { -375 BlockCacheKey cacheKey = new BlockCacheKey(this.getFileContext().getHFileName(), -376 block.getOffset(), this.isPrimaryReplicaReader(), block.getBlockType()); -377 blockCache.returnBlock(cacheKey, block); -378} -379 } -380 /** -381 * @return the first key in the file. May be null if file has no entries. Note -382 * that this is not the first row key, but rather the byte form of the -383 * first KeyValue. -384 */ -385 @Override -386 public OptionalCell getFirstKey() { -387if (dataBlockIndexReader == null) { -388 throw new BlockIndexNotLoadedException(); -389} -390return dataBlockIndexReader.isEmpty() ? Optional.empty() -391: Optional.of(dataBlockIndexReader.getRootBlockKey(0)); -392 } -393 -394 /** -395 * TODO left from {@link HFile} version 1: move this to StoreFile after Ryan's -396 * patch goes in to eliminate {@link KeyValue} here. -397 * -398 * @return the first row key, or null if the file is empty. -399 */ -400 @Override -401 public Optionalbyte[] getFirstRowKey() { -402// We have to copy the row part to form the row key alone -403return getFirstKey().map(CellUtil::cloneRow); -404 } -405 -406 /** -407 * TODO left from {@link HFile} version 1: move this to StoreFile after -408 * Ryan's patch goes in to eliminate {@link KeyValue} here. -409 * -410 * @return the last row key, or null if the file is empty. -411 */ -412 @Override -413 public Optionalbyte[] getLastRowKey() { -414// We have to copy the row part to form the row key alone -415return getLastKey().map(CellUtil::cloneRow); -416 } -417 -418 /** @return number of KV entries in this HFile */ -419 @Override -420 public long getEntries() { -421return trailer.getEntryCount(); -422 } -423 -424 /** @return comparator */ -425 @Override -426 public CellComparator getComparator() { -427return comparator; -428 } -429 -430 /** @return compression algorithm */ -431 @Override -432 public Compression.Algorithm getCompressionAlgorithm() { -433return compressAlgo; -434 } -435 -436 /** -437 * @return the total heap size of data and meta block indexes in bytes. Does -438 * not take into account non-root blocks of a multilevel data index. -439 */ -440 @Override -441 public long indexSize() { -442return (dataBlockIndexReader != null ? dataBlockIndexReader.heapSize() : 0) -443+ ((metaBlockIndexReader != null) ? metaBlockIndexReader.heapSize() -444: 0); -445 } -446 -447 @Override -448 public String getName() { -449return name; -450 } -451 -452 @Override -453 public HFileBlockIndex.BlockIndexReader getDataBlockIndexReader() { -454return dataBlockIndexReader; -455 } -456 -457 @Override -458 public FixedFileTrailer getTrailer() { -459return trailer; -460 } -461 -462 @Override -463 public boolean isPrimaryReplicaReader() { -464return primaryReplicaReader; -465 } -466 -467 @Override -468 public FileInfo loadFileInfo() throws IOException { -469return fileInfo; -470 } -471 -472 /** -473 * An exception thrown when an operation requiring a scanner to be seeked -474 * is invoked on a scanner that is not seeked. -475 */ -476 @SuppressWarnings("serial") -477 public static class NotSeekedException extends IllegalStateException { -478public NotSeekedException() { -479 super("Not seeked to a key/value"); -480} -481 } -482 -483 protected static class HFileScannerImpl implements HFileScanner { -484private ByteBuff blockBuffer; -485protected final boolean cacheBlocks; -486protected final boolean pread; -487protected final boolean isCompaction; -488private int currKeyLen; -489private int currValueLen; -490private int currMemstoreTSLen; -491private long currMemstoreTS; -492// Updated but never read? -493protected AtomicInteger blockFetches = new AtomicInteger(0); -494protected final HFile.Reader reader; -495private int currTagsLen; -496// buffer backed keyonlyKV -497private ByteBufferKeyOnlyKeyValue bufBackedKeyOnlyKv = new ByteBufferKeyOnlyKeyValue(); -498// A pair for reusing
[04/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.WorkItemHdfsDir.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.WorkItemHdfsDir.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.WorkItemHdfsDir.html index a957d31..62f81b6 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.WorkItemHdfsDir.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.WorkItemHdfsDir.html @@ -142,5192 +142,5186 @@ 134import org.apache.hadoop.hbase.wal.WAL; 135import org.apache.hadoop.hbase.wal.WALFactory; 136import org.apache.hadoop.hbase.wal.WALSplitter; -137import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; -138import org.apache.hadoop.hbase.zookeeper.ZKUtil; -139import org.apache.hadoop.hbase.zookeeper.ZKWatcher; -140import org.apache.hadoop.hbase.zookeeper.ZNodePaths; -141import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException; -142import org.apache.hadoop.ipc.RemoteException; -143import org.apache.hadoop.security.UserGroupInformation; -144import org.apache.hadoop.util.ReflectionUtils; -145import org.apache.hadoop.util.Tool; -146import org.apache.hadoop.util.ToolRunner; -147import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -148import org.apache.hbase.thirdparty.com.google.common.collect.Sets; -149import org.apache.yetus.audience.InterfaceAudience; -150import org.apache.yetus.audience.InterfaceStability; -151import org.apache.zookeeper.KeeperException; -152import org.slf4j.Logger; -153import org.slf4j.LoggerFactory; -154 -155import org.apache.hbase.thirdparty.com.google.common.base.Joiner; -156import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; -157import org.apache.hbase.thirdparty.com.google.common.collect.ImmutableList; -158import org.apache.hbase.thirdparty.com.google.common.collect.Lists; -159import org.apache.hbase.thirdparty.com.google.common.collect.Multimap; -160import org.apache.hbase.thirdparty.com.google.common.collect.Ordering; -161import org.apache.hbase.thirdparty.com.google.common.collect.TreeMultimap; -162 -163import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -164import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface; -165 -166/** -167 * HBaseFsck (hbck) is a tool for checking and repairing region consistency and -168 * table integrity problems in a corrupted HBase. This tool was written for hbase-1.x. It does not -169 * work with hbase-2.x; it can read state but is not allowed to change state; i.e. effect 'repair'. -170 * See hbck2 (HBASE-19121) for a hbck tool for hbase2. -171 * -172 * p -173 * Region consistency checks verify that hbase:meta, region deployment on region -174 * servers and the state of data in HDFS (.regioninfo files) all are in -175 * accordance. -176 * p -177 * Table integrity checks verify that all possible row keys resolve to exactly -178 * one region of a table. This means there are no individual degenerate -179 * or backwards regions; no holes between regions; and that there are no -180 * overlapping regions. -181 * p -182 * The general repair strategy works in two phases: -183 * ol -184 * li Repair Table Integrity on HDFS. (merge or fabricate regions) -185 * li Repair Region Consistency with hbase:meta and assignments -186 * /ol -187 * p -188 * For table integrity repairs, the tables' region directories are scanned -189 * for .regioninfo files. Each table's integrity is then verified. If there -190 * are any orphan regions (regions with no .regioninfo files) or holes, new -191 * regions are fabricated. Backwards regions are sidelined as well as empty -192 * degenerate (endkey==startkey) regions. If there are any overlapping regions, -193 * a new region is created and all data is merged into the new region. -194 * p -195 * Table integrity repairs deal solely with HDFS and could potentially be done -196 * offline -- the hbase region servers or master do not need to be running. -197 * This phase can eventually be used to completely reconstruct the hbase:meta table in -198 * an offline fashion. -199 * p -200 * Region consistency requires three conditions -- 1) valid .regioninfo file -201 * present in an HDFS region dir, 2) valid row with .regioninfo data in META, -202 * and 3) a region is deployed only at the regionserver that was assigned to -203 * with proper state in the master. -204 * p -205 * Region consistency repairs require hbase to be online so that hbck can -206 * contact the HBase master and region servers. The hbck#connect() method must -207 * first be called successfully. Much of the region consistency information -208 * is transient and less risky to repair. -209 * p -210 * If hbck is run from the command line, there are a handful of arguments that -211 *
[04/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/src-html/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html index 16c2238..a626878 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html @@ -228,1032 +228,1032 @@ 220RESERVED_KEYWORDS.add(IS_META_KEY); 221 } 222 -223 @InterfaceAudience.Private -224 public final static String NAMESPACE_FAMILY_INFO = "info"; -225 @InterfaceAudience.Private -226 public final static byte[] NAMESPACE_FAMILY_INFO_BYTES = Bytes.toBytes(NAMESPACE_FAMILY_INFO); +223 /** +224 * @deprecated namespace table has been folded into the ns family in meta table, do not use this +225 * any more. +226 */ 227 @InterfaceAudience.Private -228 public final static byte[] NAMESPACE_COL_DESC_BYTES = Bytes.toBytes("d"); -229 -230 /** -231 * pre -232 * Pattern that matches a coprocessor specification. Form is: -233 * {@code coprocessor jar file location '|' class name ['|' priority ['|' arguments]]} -234 * where arguments are {@code KEY '=' VALUE [,...]} -235 * For example: {@code hdfs:///foo.jar|com.foo.FooRegionObserver|1001|arg1=1,arg2=2} -236 * /pre -237 */ -238 private static final Pattern CP_HTD_ATTR_VALUE_PATTERN = -239 Pattern.compile("(^[^\\|]*)\\|([^\\|]+)\\|[\\s]*([\\d]*)[\\s]*(\\|.*)?$"); -240 -241 private static final String CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN = "[^=,]+"; -242 private static final String CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN = "[^,]+"; -243 private static final Pattern CP_HTD_ATTR_VALUE_PARAM_PATTERN = Pattern.compile( -244"(" + CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN + ")=(" + -245 CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN + "),?"); -246 private static final Pattern CP_HTD_ATTR_KEY_PATTERN = -247 Pattern.compile("^coprocessor\\$([0-9]+)$", Pattern.CASE_INSENSITIVE); -248 /** -249 * Table descriptor for namespace table -250 */ -251 // TODO We used to set CacheDataInL1 for NS table. When we have BucketCache in file mode, now the -252 // NS data goes to File mode BC only. Test how that affect the system. If too much, we have to -253 // rethink about adding back the setCacheDataInL1 for NS table. -254 public static final TableDescriptor NAMESPACE_TABLEDESC -255= TableDescriptorBuilder.newBuilder(TableName.NAMESPACE_TABLE_NAME) -256 .setColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(NAMESPACE_FAMILY_INFO_BYTES) -257// Ten is arbitrary number. Keep versions to help debugging. -258.setMaxVersions(10) -259.setInMemory(true) -260.setBlocksize(8 * 1024) -261 .setScope(HConstants.REPLICATION_SCOPE_LOCAL) -262.build()) -263 .build(); -264 private final ModifyableTableDescriptor desc; +228 @Deprecated +229 public final static String NAMESPACE_FAMILY_INFO = "info"; +230 +231 /** +232 * @deprecated namespace table has been folded into the ns family in meta table, do not use this +233 * any more. +234 */ +235 @InterfaceAudience.Private +236 @Deprecated +237 public final static byte[] NAMESPACE_FAMILY_INFO_BYTES = Bytes.toBytes(NAMESPACE_FAMILY_INFO); +238 +239 /** +240 * @deprecated namespace table has been folded into the ns family in meta table, do not use this +241 * any more. +242 */ +243 @InterfaceAudience.Private +244 @Deprecated +245 public final static byte[] NAMESPACE_COL_DESC_BYTES = Bytes.toBytes("d"); +246 +247 /** +248 * pre +249 * Pattern that matches a coprocessor specification. Form is: +250 * {@code coprocessor jar file location '|' class name ['|' priority ['|' arguments]]} +251 * where arguments are {@code KEY '=' VALUE [,...]} +252 * For example: {@code hdfs:///foo.jar|com.foo.FooRegionObserver|1001|arg1=1,arg2=2} +253 * /pre +254 */ +255 private static final Pattern CP_HTD_ATTR_VALUE_PATTERN = +256 Pattern.compile("(^[^\\|]*)\\|([^\\|]+)\\|[\\s]*([\\d]*)[\\s]*(\\|.*)?$"); +257 +258 private static final String CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN = "[^=,]+"; +259 private static final String CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN = "[^,]+"; +260 private static final Pattern CP_HTD_ATTR_VALUE_PARAM_PATTERN = Pattern.compile( +261"(" + CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN + ")=(" + +262 CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN + "),?"); +263 private static final Pattern CP_HTD_ATTR_KEY_PATTERN = +264 Pattern.compile("^coprocessor\\$([0-9]+)$", Pattern.CASE_INSENSITIVE); 265 266 /**
[04/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/testdevapidocs/src-html/org/apache/hadoop/hbase/security/access/TestAccessController.BulkLoadAccessTestAction.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/security/access/TestAccessController.BulkLoadAccessTestAction.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/security/access/TestAccessController.BulkLoadAccessTestAction.html index 5062e9b..23b4be7 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/security/access/TestAccessController.BulkLoadAccessTestAction.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/security/access/TestAccessController.BulkLoadAccessTestAction.html @@ -282,7 +282,7 @@ 274 public static void tearDownAfterClass() throws Exception { 275cleanUp(); 276TEST_UTIL.shutdownMiniCluster(); -277int total = TableAuthManager.getTotalRefCount(); +277int total = AuthManager.getTotalRefCount(); 278assertTrue("Unexpected reference count: " + total, total == 0); 279 } 280 @@ -1642,12 +1642,12 @@ 1634 } 1635 1636 UserPermission ownerperm = -1637 new UserPermission(Bytes.toBytes(USER_OWNER.getName()), tableName, null, Action.values()); +1637 new UserPermission(USER_OWNER.getName(), tableName, Action.values()); 1638 assertTrue("Owner should have all permissions on table", 1639 hasFoundUserPermission(ownerperm, perms)); 1640 1641 User user = User.createUserForTesting(TEST_UTIL.getConfiguration(), "user", new String[0]); -1642 byte[] userName = Bytes.toBytes(user.getShortName()); +1642 String userName = user.getShortName(); 1643 1644 UserPermission up = 1645 new UserPermission(userName, tableName, family1, qualifier, Permission.Action.READ); @@ -1733,7 +1733,7 @@ 1725 } 1726 1727 UserPermission newOwnerperm = -1728 new UserPermission(Bytes.toBytes(newOwner.getName()), tableName, null, Action.values()); +1728 new UserPermission(newOwner.getName(), tableName, Action.values()); 1729 assertTrue("New owner should have all permissions on table", 1730 hasFoundUserPermission(newOwnerperm, perms)); 1731} finally { @@ -1757,1888 +1757,1898 @@ 1749 1750CollectionString superUsers = Superusers.getSuperUsers(); 1751ListUserPermission adminPerms = new ArrayList(superUsers.size() + 1); -1752adminPerms.add(new UserPermission(Bytes.toBytes(USER_ADMIN.getShortName()), -1753 AccessControlLists.ACL_TABLE_NAME, null, null, Bytes.toBytes("ACRW"))); -1754 -1755for(String user: superUsers) { -1756 adminPerms.add(new UserPermission(Bytes.toBytes(user), AccessControlLists.ACL_TABLE_NAME, -1757 null, null, Action.values())); -1758} -1759assertTrue("Only super users, global users and user admin has permission on table hbase:acl " + -1760"per setup", perms.size() == 5 + superUsers.size() -1761 hasFoundUserPermission(adminPerms, perms)); -1762 } -1763 -1764 /** global operations */ -1765 private void verifyGlobal(AccessTestAction action) throws Exception { -1766verifyAllowed(action, SUPERUSER); -1767 -1768verifyDenied(action, USER_CREATE, USER_RW, USER_NONE, USER_RO); -1769 } -1770 -1771 @Test -1772 public void testCheckPermissions() throws Exception { -1773// -- -1774// test global permissions -1775AccessTestAction globalAdmin = new AccessTestAction() { -1776 @Override -1777 public Void run() throws Exception { -1778checkGlobalPerms(TEST_UTIL, Permission.Action.ADMIN); -1779return null; -1780 } -1781}; -1782// verify that only superuser can admin -1783verifyGlobal(globalAdmin); -1784 -1785// -- -1786// test multiple permissions -1787AccessTestAction globalReadWrite = new AccessTestAction() { -1788 @Override -1789 public Void run() throws Exception { -1790checkGlobalPerms(TEST_UTIL, Permission.Action.READ, Permission.Action.WRITE); -1791return null; -1792 } -1793}; +1752adminPerms.add(new UserPermission(USER_ADMIN.getShortName(), Bytes.toBytes("ACRW"))); +1753for(String user: superUsers) { +1754 // Global permission +1755 adminPerms.add(new UserPermission(user, Action.values())); +1756} +1757assertTrue("Only super users, global users and user admin has permission on table hbase:acl " + +1758"per setup", perms.size() == 5 + superUsers.size() +1759 hasFoundUserPermission(adminPerms, perms)); +1760 } +1761 +1762 /** global operations */ +1763 private void verifyGlobal(AccessTestAction action) throws Exception { +1764verifyAllowed(action, SUPERUSER); +1765 +1766verifyDenied(action, USER_CREATE, USER_RW, USER_NONE, USER_RO); +1767 } +1768 +1769 @Test
[04/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestMasterProcedureScheduler.TestPeerProcedure.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestMasterProcedureScheduler.TestPeerProcedure.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestMasterProcedureScheduler.TestPeerProcedure.html index c9e0e55..f1b9105 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestMasterProcedureScheduler.TestPeerProcedure.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestMasterProcedureScheduler.TestPeerProcedure.html @@ -37,157 +37,157 @@ 029import java.util.Arrays; 030import java.util.List; 031import org.apache.hadoop.hbase.HBaseClassTestRule; -032import org.apache.hadoop.hbase.HRegionInfo; -033import org.apache.hadoop.hbase.ServerName; -034import org.apache.hadoop.hbase.TableName; -035import org.apache.hadoop.hbase.client.RegionInfo; -036import org.apache.hadoop.hbase.client.RegionInfoBuilder; -037import org.apache.hadoop.hbase.master.locking.LockProcedure; -038import org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType; -039import org.apache.hadoop.hbase.procedure2.LockType; -040import org.apache.hadoop.hbase.procedure2.LockedResource; -041import org.apache.hadoop.hbase.procedure2.LockedResourceType; -042import org.apache.hadoop.hbase.procedure2.Procedure; -043import org.apache.hadoop.hbase.procedure2.ProcedureEvent; -044import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility.TestProcedure; -045import org.apache.hadoop.hbase.testclassification.MasterTests; -046import org.apache.hadoop.hbase.testclassification.SmallTests; -047import org.apache.hadoop.hbase.util.Bytes; -048import org.junit.After; -049import org.junit.Before; -050import org.junit.ClassRule; -051import org.junit.Rule; -052import org.junit.Test; -053import org.junit.experimental.categories.Category; -054import org.junit.rules.TestName; -055import org.slf4j.Logger; -056import org.slf4j.LoggerFactory; -057 -058@Category({MasterTests.class, SmallTests.class}) -059public class TestMasterProcedureScheduler { -060 -061 @ClassRule -062 public static final HBaseClassTestRule CLASS_RULE = -063 HBaseClassTestRule.forClass(TestMasterProcedureScheduler.class); -064 -065 private static final Logger LOG = LoggerFactory.getLogger(TestMasterProcedureScheduler.class); -066 -067 private MasterProcedureScheduler queue; -068 -069 @Rule -070 public TestName name = new TestName(); -071 -072 @Before -073 public void setUp() throws IOException { -074queue = new MasterProcedureScheduler(); -075queue.start(); -076 } -077 -078 @After -079 public void tearDown() throws IOException { -080assertEquals("proc-queue expected to be empty", 0, queue.size()); -081queue.stop(); -082queue.clear(); -083 } -084 -085 /** -086 * Verify simple create/insert/fetch/delete of the table queue. -087 */ -088 @Test -089 public void testSimpleTableOpsQueues() throws Exception { -090final int NUM_TABLES = 10; -091final int NUM_ITEMS = 10; -092 -093int count = 0; -094for (int i = 1; i = NUM_TABLES; ++i) { -095 TableName tableName = TableName.valueOf(String.format("test-%04d", i)); -096 // insert items -097 for (int j = 1; j = NUM_ITEMS; ++j) { -098queue.addBack(new TestTableProcedure(i * 1000 + j, tableName, -099 TableProcedureInterface.TableOperationType.EDIT)); -100assertEquals(++count, queue.size()); -101 } -102} -103assertEquals(NUM_TABLES * NUM_ITEMS, queue.size()); -104 -105for (int j = 1; j = NUM_ITEMS; ++j) { -106 for (int i = 1; i = NUM_TABLES; ++i) { -107Procedure proc = queue.poll(); -108assertTrue(proc != null); -109TableName tableName = ((TestTableProcedure)proc).getTableName(); -110 queue.waitTableExclusiveLock(proc, tableName); -111 queue.wakeTableExclusiveLock(proc, tableName); -112queue.completionCleanup(proc); -113assertEquals(--count, queue.size()); -114assertEquals(i * 1000 + j, proc.getProcId()); -115 } -116} -117assertEquals(0, queue.size()); -118 -119for (int i = 1; i = NUM_TABLES; ++i) { -120 final TableName tableName = TableName.valueOf(String.format("test-%04d", i)); -121 final TestTableProcedure dummyProc = new TestTableProcedure(100, tableName, -122 TableProcedureInterface.TableOperationType.DELETE); -123 // complete the table deletion -124 assertTrue(queue.markTableAsDeleted(tableName, dummyProc)); -125} -126 } -127 -128 /** -129 * Check that the table queue is not deletable until every procedure -130 * in-progress is completed (this is a special case for write-locks). -131 */
[04/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.html index 9b964f6..98ef11a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.html @@ -105,7 +105,7 @@ 097 * will first be initialized to the oldest file's tracker(which is stored in the trailer), using the 098 * method {@link ProcedureStoreTracker#resetTo(ProcedureStoreTracker, boolean)}, and then merge it 099 * with the tracker of every newer wal files, using the -100 * {@link ProcedureStoreTracker#setDeletedIfModifiedInBoth(ProcedureStoreTracker, boolean)}. +100 * {@link ProcedureStoreTracker#setDeletedIfModifiedInBoth(ProcedureStoreTracker)}. 101 * If we find out 102 * that all the modified procedures for the oldest wal file are modified or deleted in newer wal 103 * files, then we can delete it. This is because that, every time we call @@ -1181,244 +1181,243 @@ 1173} 1174 1175// compute the holding tracker. -1176// - the first WAL is used for the 'updates' -1177// - the global tracker is passed in first to decide which procedures are not -1178//exist anymore, so we can mark them as deleted in holdingCleanupTracker. -1179//Only global tracker have the whole picture here. -1180// - the other WALs are scanned to remove procs already updated in a newer wal. -1181//If it is updated in a newer wal, we can mark it as delelted in holdingCleanupTracker -1182//But, we can not delete it if it was shown deleted in the newer wal, as said -1183//above. -1184// TODO: exit early if holdingCleanupTracker.isEmpty() -1185 holdingCleanupTracker.resetTo(logs.getFirst().getTracker(), true); -1186//Passing in the global tracker, we can delete the procedures not in the global -1187//tracker, because they are deleted in the later logs -1188 holdingCleanupTracker.setDeletedIfModifiedInBoth(storeTracker, true); -1189for (int i = 1, size = logs.size() - 1; i size; ++i) { -1190 // Set deleteIfNotExists to false since a single log's tracker is passed in. -1191 // Since a specific procedure may not show up in the log at all(not executed or -1192 // updated during the time), we can not delete the procedure just because this log -1193 // don't have the info of the procedure. We can delete the procedure only if -1194 // in this log's tracker, it was cleanly showed that the procedure is modified or deleted -1195 // in the corresponding BitSetNode. -1196 holdingCleanupTracker.setDeletedIfModifiedInBoth(logs.get(i).getTracker(), false); -1197} -1198 } -1199 -1200 /** -1201 * Remove all logs with logId = {@code lastLogId}. -1202 */ -1203 private void removeAllLogs(long lastLogId, String why) { -1204if (logs.size() = 1) { -1205 return; -1206} -1207 -1208LOG.info("Remove all state logs with ID less than {}, since {}", lastLogId, why); -1209 -1210boolean removed = false; -1211while (logs.size() 1) { -1212 ProcedureWALFile log = logs.getFirst(); -1213 if (lastLogId log.getLogId()) { -1214break; -1215 } -1216 removeLogFile(log, walArchiveDir); -1217 removed = true; -1218} -1219 -1220if (removed) { -1221 buildHoldingCleanupTracker(); -1222} -1223 } -1224 -1225 private boolean removeLogFile(final ProcedureWALFile log, final Path walArchiveDir) { -1226try { -1227 LOG.trace("Removing log={}", log); -1228 log.removeFile(walArchiveDir); -1229 logs.remove(log); -1230 LOG.debug("Removed log={}, activeLogs={}", log, logs); -1231 assert logs.size() 0 : "expected at least one log"; -1232} catch (IOException e) { -1233 LOG.error("Unable to remove log: " + log, e); -1234 return false; -1235} -1236return true; -1237 } -1238 -1239 // == -1240 // FileSystem Log Files helpers -1241 // == -1242 public Path getWALDir() { -1243return this.walDir; -1244 } -1245 -1246 @VisibleForTesting -1247 Path getWalArchiveDir() { -1248return this.walArchiveDir; -1249 } -1250 -1251 public FileSystem getFileSystem() { -1252return this.fs; -1253 } -1254 -1255 protected Path getLogFilePath(final long logId) throws IOException { -1256return new Path(walDir, String.format(LOG_PREFIX + "%020d.log", logId)); -1257 } -1258 -1259 private static long
[04/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.HStoreForTesting.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.HStoreForTesting.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.HStoreForTesting.html index ed3db7a..156dabb 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.HStoreForTesting.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/regionserver/TestHRegion.HStoreForTesting.html @@ -5542,785 +5542,825 @@ 5534 } 5535 5536 @Test -5537 public void testWriteRequestsCounter() throws IOException { -5538byte[] fam = Bytes.toBytes("info"); -5539byte[][] families = { fam }; -5540this.region = initHRegion(tableName, method, CONF, families); +5537 public void testReverseScanWhenPutCellsAfterOpenReverseScan() throws Exception { +5538byte[] cf1 = Bytes.toBytes("CF1"); +5539byte[][] families = { cf1 }; +5540byte[] col = Bytes.toBytes("C"); 5541 -5542Assert.assertEquals(0L, region.getWriteRequestsCount()); -5543 -5544Put put = new Put(row); -5545put.addColumn(fam, fam, fam); -5546 -5547Assert.assertEquals(0L, region.getWriteRequestsCount()); -5548region.put(put); -5549Assert.assertEquals(1L, region.getWriteRequestsCount()); -5550region.put(put); -5551Assert.assertEquals(2L, region.getWriteRequestsCount()); -5552region.put(put); -5553Assert.assertEquals(3L, region.getWriteRequestsCount()); -5554 -region.delete(new Delete(row)); -5556Assert.assertEquals(4L, region.getWriteRequestsCount()); -5557 } -5558 -5559 @Test -5560 public void testOpenRegionWrittenToWAL() throws Exception { -5561final ServerName serverName = ServerName.valueOf(name.getMethodName(), 100, 42); -5562final RegionServerServices rss = spy(TEST_UTIL.createMockRegionServerService(serverName)); -5563 -5564HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(name.getMethodName())); -5565htd.addFamily(new HColumnDescriptor(fam1)); -5566htd.addFamily(new HColumnDescriptor(fam2)); -5567 -5568HRegionInfo hri = new HRegionInfo(htd.getTableName(), -5569 HConstants.EMPTY_BYTE_ARRAY, HConstants.EMPTY_BYTE_ARRAY); -5570 -5571// open the region w/o rss and wal and flush some files -5572region = -5573 HBaseTestingUtility.createRegionAndWAL(hri, TEST_UTIL.getDataTestDir(), TEST_UTIL -5574 .getConfiguration(), htd); -5575assertNotNull(region); -5576 -5577// create a file in fam1 for the region before opening in OpenRegionHandler -5578region.put(new Put(Bytes.toBytes("a")).addColumn(fam1, fam1, fam1)); -5579region.flush(true); -5580 HBaseTestingUtility.closeRegionAndWAL(region); +5542HBaseConfiguration conf = new HBaseConfiguration(); +5543this.region = initHRegion(tableName, method, conf, families); +5544 +5545Put put = new Put(Bytes.toBytes("16")); +5546put.addColumn(cf1, col, Bytes.toBytes("val")); +5547region.put(put); +5548Put put2 = new Put(Bytes.toBytes("15")); +5549put2.addColumn(cf1, col, Bytes.toBytes("val")); +5550region.put(put2); +5551 +5552// Create a reverse scan +5553Scan scan = new Scan(Bytes.toBytes("16")); +5554scan.setReversed(true); +RegionScannerImpl scanner = region.getScanner(scan); +5556 +5557// Put a lot of cells that have sequenceIDs grater than the readPt of the reverse scan +5558for (int i = 10; i 20; i++) { +5559 Put p = new Put(Bytes.toBytes("" + i)); +5560 p.addColumn(cf1, col, Bytes.toBytes("" + i)); +5561 region.put(p); +5562} +5563ListCell currRow = new ArrayList(); +5564boolean hasNext; +5565do { +5566 hasNext = scanner.next(currRow); +5567} while (hasNext); +5568 +5569assertEquals(2, currRow.size()); +5570assertEquals("16", Bytes.toString(currRow.get(0).getRowArray(), +5571 currRow.get(0).getRowOffset(), currRow.get(0).getRowLength())); +5572assertEquals("15", Bytes.toString(currRow.get(1).getRowArray(), +5573 currRow.get(1).getRowOffset(), currRow.get(1).getRowLength())); +5574 } +5575 +5576 @Test +5577 public void testWriteRequestsCounter() throws IOException { +5578byte[] fam = Bytes.toBytes("info"); +5579byte[][] families = { fam }; +5580this.region = initHRegion(tableName, method, CONF, families); 5581 -5582ArgumentCaptorWALEdit editCaptor = ArgumentCaptor.forClass(WALEdit.class); +5582Assert.assertEquals(0L, region.getWriteRequestsCount()); 5583 -5584// capture append() calls -5585WAL wal = mockWAL(); -5586when(rss.getWAL((HRegionInfo) any())).thenReturn(wal); -5587 -5588region = HRegion.openHRegion(hri, htd, rss.getWAL(hri), -5589
[04/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/util/class-use/ByteBufferUtils.Comparer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/class-use/ByteBufferUtils.Comparer.html b/devapidocs/org/apache/hadoop/hbase/util/class-use/ByteBufferUtils.Comparer.html new file mode 100644 index 000..d3d202d --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/util/class-use/ByteBufferUtils.Comparer.html @@ -0,0 +1,195 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +Uses of Class org.apache.hadoop.hbase.util.ByteBufferUtils.Comparer (Apache HBase 3.0.0-SNAPSHOT API) + + + + + + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + + +Uses of Classorg.apache.hadoop.hbase.util.ByteBufferUtils.Comparer + + + + + +Packages that use ByteBufferUtils.Comparer + +Package +Description + + + +org.apache.hadoop.hbase.util + + + + + + + + + + +Uses of ByteBufferUtils.Comparer in org.apache.hadoop.hbase.util + +Subclasses of ByteBufferUtils.Comparer in org.apache.hadoop.hbase.util + +Modifier and Type +Class and Description + + + +(package private) static class +ByteBufferUtils.ComparerHolder.PureJavaComparer + + +(package private) static class +ByteBufferUtils.ComparerHolder.UnsafeComparer + + + + +Fields in org.apache.hadoop.hbase.util declared as ByteBufferUtils.Comparer + +Modifier and Type +Field and Description + + + +(package private) static ByteBufferUtils.Comparer +ByteBufferUtils.ComparerHolder.BEST_COMPARER + + + + +Methods in org.apache.hadoop.hbase.util that return ByteBufferUtils.Comparer + +Modifier and Type +Method and Description + + + +(package private) static ByteBufferUtils.Comparer +ByteBufferUtils.ComparerHolder.getBestComparer() + + + + + + + + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/util/class-use/ByteBufferUtils.ComparerHolder.PureJavaComparer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/class-use/ByteBufferUtils.ComparerHolder.PureJavaComparer.html b/devapidocs/org/apache/hadoop/hbase/util/class-use/ByteBufferUtils.ComparerHolder.PureJavaComparer.html new file mode 100644 index 000..6e8eb4d --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/util/class-use/ByteBufferUtils.ComparerHolder.PureJavaComparer.html @@ -0,0 +1,165 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +Uses of Class org.apache.hadoop.hbase.util.ByteBufferUtils.ComparerHolder.PureJavaComparer (Apache HBase 3.0.0-SNAPSHOT API) + + + + + + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + + +Uses of Classorg.apache.hadoop.hbase.util.ByteBufferUtils.ComparerHolder.PureJavaComparer + + + + + +Packages that use ByteBufferUtils.ComparerHolder.PureJavaComparer + +Package +Description + + + +org.apache.hadoop.hbase.util + + + + + + + + + + +Uses of ByteBufferUtils.ComparerHolder.PureJavaComparer in org.apache.hadoop.hbase.util + +Fields in org.apache.hadoop.hbase.util declared as ByteBufferUtils.ComparerHolder.PureJavaComparer + +Modifier
[04/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestModifyTableProcedure.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestModifyTableProcedure.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestModifyTableProcedure.html index ad8e4a5..ec3e399 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestModifyTableProcedure.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/procedure/TestModifyTableProcedure.html @@ -30,385 +30,383 @@ 022import static org.junit.Assert.assertTrue; 023 024import java.io.IOException; -025 -026import org.apache.hadoop.hbase.DoNotRetryIOException; -027import org.apache.hadoop.hbase.HBaseClassTestRule; -028import org.apache.hadoop.hbase.HColumnDescriptor; -029import org.apache.hadoop.hbase.HTableDescriptor; -030import org.apache.hadoop.hbase.InvalidFamilyOperationException; -031import org.apache.hadoop.hbase.TableName; -032import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -033import org.apache.hadoop.hbase.client.PerClientRandomNonceGenerator; -034import org.apache.hadoop.hbase.client.RegionInfo; -035import org.apache.hadoop.hbase.client.TableDescriptor; -036import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -037import org.apache.hadoop.hbase.master.procedure.MasterProcedureTestingUtility.StepHook; -038import org.apache.hadoop.hbase.procedure2.Procedure; -039import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; -040import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility; -041import org.apache.hadoop.hbase.testclassification.MasterTests; -042import org.apache.hadoop.hbase.testclassification.MediumTests; -043import org.apache.hadoop.hbase.util.Bytes; -044import org.apache.hadoop.hbase.util.NonceKey; -045import org.junit.Assert; -046import org.junit.ClassRule; -047import org.junit.Rule; -048import org.junit.Test; -049import org.junit.experimental.categories.Category; -050import org.junit.rules.TestName; -051 -052@Category({MasterTests.class, MediumTests.class}) -053public class TestModifyTableProcedure extends TestTableDDLProcedureBase { -054 -055 @ClassRule -056 public static final HBaseClassTestRule CLASS_RULE = -057 HBaseClassTestRule.forClass(TestModifyTableProcedure.class); -058 -059 @Rule public TestName name = new TestName(); -060 -061 @Test -062 public void testModifyTable() throws Exception { -063final TableName tableName = TableName.valueOf(name.getMethodName()); -064final ProcedureExecutorMasterProcedureEnv procExec = getMasterProcedureExecutor(); -065 -066 MasterProcedureTestingUtility.createTable(procExec, tableName, null, "cf"); -067 UTIL.getAdmin().disableTable(tableName); -068 -069// Modify the table descriptor -070HTableDescriptor htd = new HTableDescriptor(UTIL.getAdmin().getTableDescriptor(tableName)); -071 -072// Test 1: Modify 1 property -073long newMaxFileSize = htd.getMaxFileSize() * 2; -074htd.setMaxFileSize(newMaxFileSize); -075htd.setRegionReplication(3); -076 -077long procId1 = ProcedureTestingUtility.submitAndWait( -078procExec, new ModifyTableProcedure(procExec.getEnvironment(), htd)); -079 ProcedureTestingUtility.assertProcNotFailed(procExec.getResult(procId1)); -080 -081HTableDescriptor currentHtd = UTIL.getAdmin().getTableDescriptor(tableName); -082assertEquals(newMaxFileSize, currentHtd.getMaxFileSize()); -083 -084// Test 2: Modify multiple properties -085boolean newReadOnlyOption = htd.isReadOnly() ? false : true; -086long newMemStoreFlushSize = htd.getMemStoreFlushSize() * 2; -087htd.setReadOnly(newReadOnlyOption); -088 htd.setMemStoreFlushSize(newMemStoreFlushSize); -089 -090long procId2 = ProcedureTestingUtility.submitAndWait( -091procExec, new ModifyTableProcedure(procExec.getEnvironment(), htd)); -092 ProcedureTestingUtility.assertProcNotFailed(procExec.getResult(procId2)); -093 -094currentHtd = UTIL.getAdmin().getTableDescriptor(tableName); -095assertEquals(newReadOnlyOption, currentHtd.isReadOnly()); -096assertEquals(newMemStoreFlushSize, currentHtd.getMemStoreFlushSize()); -097 } -098 -099 @Test -100 public void testModifyTableAddCF() throws Exception { -101final TableName tableName = TableName.valueOf(name.getMethodName()); -102final ProcedureExecutorMasterProcedureEnv procExec = getMasterProcedureExecutor(); -103 -104 MasterProcedureTestingUtility.createTable(procExec, tableName, null, "cf1"); -105HTableDescriptor currentHtd = UTIL.getAdmin().getTableDescriptor(tableName); -106assertEquals(1, currentHtd.getFamiliesKeys().size()); -107 -108// Test 1: Modify the table descriptor online -109String cf2 = "cf2"; -110HTableDescriptor htd =
[04/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/license.html -- diff --git a/license.html b/license.html index 22567b5..252581d 100644 --- a/license.html +++ b/license.html @@ -7,7 +7,7 @@ - + Apache HBase Project Licenses @@ -491,7 +491,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-10-11 + Last Published: 2018-10-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/mail-lists.html -- diff --git a/mail-lists.html b/mail-lists.html index 2fee6c0..34eaf86 100644 --- a/mail-lists.html +++ b/mail-lists.html @@ -7,7 +7,7 @@ - + Apache HBase Project Mailing Lists @@ -341,7 +341,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-10-11 + Last Published: 2018-10-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/metrics.html -- diff --git a/metrics.html b/metrics.html index 0938acd..c44190a 100644 --- a/metrics.html +++ b/metrics.html @@ -7,7 +7,7 @@ - + Apache HBase Apache HBase (TM) Metrics @@ -459,7 +459,7 @@ export HBASE_REGIONSERVER_OPTS=$HBASE_JMX_OPTS -Dcom.sun.management.jmxrem https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-10-11 + Last Published: 2018-10-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/old_news.html -- diff --git a/old_news.html b/old_news.html index 45acdd2..ce46430 100644 --- a/old_news.html +++ b/old_news.html @@ -7,7 +7,7 @@ - + Apache HBase Old Apache HBase (TM) News @@ -440,7 +440,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-10-11 + Last Published: 2018-10-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/plugin-management.html -- diff --git a/plugin-management.html b/plugin-management.html index a745e35..a229b91 100644 --- a/plugin-management.html +++ b/plugin-management.html @@ -7,7 +7,7 @@ - + Apache HBase Project Plugin Management @@ -440,7 +440,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-10-11 + Last Published: 2018-10-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/plugins.html -- diff --git a/plugins.html b/plugins.html index f87797f..290350a 100644 --- a/plugins.html +++ b/plugins.html @@ -7,7 +7,7 @@ - + Apache HBase Project Plugins @@ -375,7 +375,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-10-11 + Last Published: 2018-10-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/poweredbyhbase.html -- diff --git a/poweredbyhbase.html b/poweredbyhbase.html index 6016ce6..a614e2a 100644 --- a/poweredbyhbase.html +++ b/poweredbyhbase.html @@ -7,7 +7,7 @@ - + Apache HBase Powered By Apache HBase @@ -769,7 +769,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-10-11 + Last Published: 2018-10-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/project-info.html -- diff --git a/project-info.html b/project-info.html index
[04/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/testdevapidocs/org/apache/hadoop/hbase/procedure2/class-use/TestProcedureSkipPersistence.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/procedure2/class-use/TestProcedureSkipPersistence.html b/testdevapidocs/org/apache/hadoop/hbase/procedure2/class-use/TestProcedureSkipPersistence.html new file mode 100644 index 000..e4812fc --- /dev/null +++ b/testdevapidocs/org/apache/hadoop/hbase/procedure2/class-use/TestProcedureSkipPersistence.html @@ -0,0 +1,125 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +Uses of Class org.apache.hadoop.hbase.procedure2.TestProcedureSkipPersistence (Apache HBase 3.0.0-SNAPSHOT Test API) + + + + + + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + + +Uses of Classorg.apache.hadoop.hbase.procedure2.TestProcedureSkipPersistence + +No usage of org.apache.hadoop.hbase.procedure2.TestProcedureSkipPersistence + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-frame.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-frame.html b/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-frame.html index ec42355..93c7697 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-frame.html +++ b/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-frame.html @@ -60,6 +60,8 @@ TestProcedureReplayOrder.TestTwoStepProcedure TestProcedureSchedulerConcurrency TestProcedureSchedulerConcurrency.TestProcedureWithEvent +TestProcedureSkipPersistence +TestProcedureSkipPersistence.TestProcedure TestProcedureSuspended TestProcedureSuspended.TestLockProcedure TestProcedureSuspended.TestProcEnv http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-summary.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-summary.html b/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-summary.html index 689909e..6d9932a 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-summary.html +++ b/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-summary.html @@ -280,94 +280,102 @@ -TestProcedureSuspended +TestProcedureSkipPersistence + + + +TestProcedureSkipPersistence.TestProcedure -TestProcedureSuspended.TestLockProcedure +TestProcedureSuspended -TestProcedureSuspended.TestProcEnv +TestProcedureSuspended.TestLockProcedure -TestProcedureToString +TestProcedureSuspended.TestProcEnv +TestProcedureToString + + + TestProcedureToString.BasicProcedure A do-nothing basic procedure just for testing toString. - + TestProcedureToString.BasicProcedureEnv A do-nothing environment for BasicProcedure. - + TestProcedureToString.DoublingStateStringBasicProcedure A do-nothing basic procedure that overrides the toStringState method. - + TestProcedureToString.SimpleStateMachineProcedure Do-nothing SimpleMachineProcedure for checking its toString. - + TestProcedureUtil - + TestProcedureUtil.TestProcedureNoDefaultConstructor - + TestStateMachineProcedure - + TestStateMachineProcedure.SimpleChildProcedure - + TestStateMachineProcedure.TestProcEnv - + TestStateMachineProcedure.TestSMProcedure - + TestStateMachineProcedure.TestSMProcedureBadRollback - + TestYieldProcedures - + TestYieldProcedures.TestProcEnv - + TestYieldProcedures.TestScheduler - + TestYieldProcedures.TestStateMachineProcedure - +
[04/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html index 566f410..da040ad 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html @@ -341,8361 +341,8425 @@ 333 private final int rowLockWaitDuration; 334 static final int DEFAULT_ROWLOCK_WAIT_DURATION = 3; 335 -336 // The internal wait duration to acquire a lock before read/update -337 // from the region. It is not per row. The purpose of this wait time -338 // is to avoid waiting a long time while the region is busy, so that -339 // we can release the IPC handler soon enough to improve the -340 // availability of the region server. It can be adjusted by -341 // tuning configuration "hbase.busy.wait.duration". -342 final long busyWaitDuration; -343 static final long DEFAULT_BUSY_WAIT_DURATION = HConstants.DEFAULT_HBASE_RPC_TIMEOUT; -344 -345 // If updating multiple rows in one call, wait longer, -346 // i.e. waiting for busyWaitDuration * # of rows. However, -347 // we can limit the max multiplier. -348 final int maxBusyWaitMultiplier; -349 -350 // Max busy wait duration. There is no point to wait longer than the RPC -351 // purge timeout, when a RPC call will be terminated by the RPC engine. -352 final long maxBusyWaitDuration; -353 -354 // Max cell size. If nonzero, the maximum allowed size for any given cell -355 // in bytes -356 final long maxCellSize; -357 -358 // Number of mutations for minibatch processing. -359 private final int miniBatchSize; +336 private Path regionDir; +337 private FileSystem walFS; +338 +339 // The internal wait duration to acquire a lock before read/update +340 // from the region. It is not per row. The purpose of this wait time +341 // is to avoid waiting a long time while the region is busy, so that +342 // we can release the IPC handler soon enough to improve the +343 // availability of the region server. It can be adjusted by +344 // tuning configuration "hbase.busy.wait.duration". +345 final long busyWaitDuration; +346 static final long DEFAULT_BUSY_WAIT_DURATION = HConstants.DEFAULT_HBASE_RPC_TIMEOUT; +347 +348 // If updating multiple rows in one call, wait longer, +349 // i.e. waiting for busyWaitDuration * # of rows. However, +350 // we can limit the max multiplier. +351 final int maxBusyWaitMultiplier; +352 +353 // Max busy wait duration. There is no point to wait longer than the RPC +354 // purge timeout, when a RPC call will be terminated by the RPC engine. +355 final long maxBusyWaitDuration; +356 +357 // Max cell size. If nonzero, the maximum allowed size for any given cell +358 // in bytes +359 final long maxCellSize; 360 -361 // negative number indicates infinite timeout -362 static final long DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L; -363 final ExecutorService rowProcessorExecutor = Executors.newCachedThreadPool(); -364 -365 private final ConcurrentHashMapRegionScanner, Long scannerReadPoints; -366 -367 /** -368 * The sequence ID that was enLongAddered when this region was opened. -369 */ -370 private long openSeqNum = HConstants.NO_SEQNUM; -371 -372 /** -373 * The default setting for whether to enable on-demand CF loading for -374 * scan requests to this region. Requests can override it. -375 */ -376 private boolean isLoadingCfsOnDemandDefault = false; -377 -378 private final AtomicInteger majorInProgress = new AtomicInteger(0); -379 private final AtomicInteger minorInProgress = new AtomicInteger(0); +361 // Number of mutations for minibatch processing. +362 private final int miniBatchSize; +363 +364 // negative number indicates infinite timeout +365 static final long DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L; +366 final ExecutorService rowProcessorExecutor = Executors.newCachedThreadPool(); +367 +368 private final ConcurrentHashMapRegionScanner, Long scannerReadPoints; +369 +370 /** +371 * The sequence ID that was enLongAddered when this region was opened. +372 */ +373 private long openSeqNum = HConstants.NO_SEQNUM; +374 +375 /** +376 * The default setting for whether to enable on-demand CF loading for +377 * scan requests to this region. Requests can override it. +378 */ +379 private boolean isLoadingCfsOnDemandDefault = false; 380 -381 // -382 // Context: During replay we want to ensure that we do not lose any data. So, we -383 // have to be conservative in how we replay wals. For each store, we calculate -384 // the maxSeqId up to which the store was flushed.
[04/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html index 4a11f27..7c7966d 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.PostOpenDeployTasksThread.html @@ -49,287 +49,290 @@ 041 * Handles opening of a region on a region server. 042 * p 043 * This is executed after receiving an OPEN RPC from the master or client. -044 */ -045@InterfaceAudience.Private -046public class OpenRegionHandler extends EventHandler { -047 private static final Logger LOG = LoggerFactory.getLogger(OpenRegionHandler.class); -048 -049 protected final RegionServerServices rsServices; -050 -051 private final RegionInfo regionInfo; -052 private final TableDescriptor htd; -053 private final long masterSystemTime; -054 -055 public OpenRegionHandler(final Server server, -056 final RegionServerServices rsServices, RegionInfo regionInfo, -057 TableDescriptor htd, long masterSystemTime) { -058this(server, rsServices, regionInfo, htd, masterSystemTime, EventType.M_RS_OPEN_REGION); -059 } -060 -061 protected OpenRegionHandler(final Server server, -062 final RegionServerServices rsServices, final RegionInfo regionInfo, -063 final TableDescriptor htd, long masterSystemTime, EventType eventType) { -064super(server, eventType); -065this.rsServices = rsServices; -066this.regionInfo = regionInfo; -067this.htd = htd; -068this.masterSystemTime = masterSystemTime; -069 } -070 -071 public RegionInfo getRegionInfo() { -072return regionInfo; -073 } -074 -075 @Override -076 public void process() throws IOException { -077boolean openSuccessful = false; -078final String regionName = regionInfo.getRegionNameAsString(); -079HRegion region = null; -080 -081try { -082 if (this.server.isStopped() || this.rsServices.isStopping()) { -083return; -084 } -085 final String encodedName = regionInfo.getEncodedName(); -086 -087 // 2 different difficult situations can occur -088 // 1) The opening was cancelled. This is an expected situation -089 // 2) The region is now marked as online while we're suppose to open. This would be a bug. -090 -091 // Check that this region is not already online -092 if (this.rsServices.getRegion(encodedName) != null) { -093LOG.error("Region " + encodedName + -094" was already online when we started processing the opening. " + -095"Marking this new attempt as failed"); -096return; -097 } -098 -099 // Check that we're still supposed to open the region. -100 // If fails, just return. Someone stole the region from under us. -101 if (!isRegionStillOpening()){ -102LOG.error("Region " + encodedName + " opening cancelled"); -103return; -104 } -105 -106 // Open region. After a successful open, failures in subsequent -107 // processing needs to do a close as part of cleanup. -108 region = openRegion(); -109 if (region == null) { -110return; -111 } -112 -113 if (!updateMeta(region, masterSystemTime) || this.server.isStopped() || -114 this.rsServices.isStopping()) { -115return; -116 } -117 -118 if (!isRegionStillOpening()) { -119return; -120 } -121 -122 // Successful region open, and add it to MutableOnlineRegions -123 this.rsServices.addRegion(region); -124 openSuccessful = true; -125 -126 // Done! Successful region open -127 LOG.debug("Opened " + regionName + " on " + this.server.getServerName()); -128} finally { -129 // Do all clean up here -130 if (!openSuccessful) { -131doCleanUpOnFailedOpen(region); -132 } -133 final Boolean current = this.rsServices.getRegionsInTransitionInRS(). -134 remove(this.regionInfo.getEncodedNameAsBytes()); -135 -136 // Let's check if we have met a race condition on open cancellation -137 // A better solution would be to not have any race condition. -138 // this.rsServices.getRegionsInTransitionInRS().remove( -139 // this.regionInfo.getEncodedNameAsBytes(), Boolean.TRUE); -140 // would help. -141 if (openSuccessful) { -142if (current == null) { // Should NEVER happen, but let's be paranoid. -143 LOG.error("Bad state: we've just
[04/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html index 2c14c50..43c66a8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html @@ -46,2104 +46,2113 @@ 038import java.util.concurrent.atomic.AtomicLong; 039import java.util.stream.Collectors; 040import java.util.stream.Stream; -041import org.apache.hadoop.conf.Configuration; -042import org.apache.hadoop.hbase.HConstants; -043import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; -044import org.apache.hadoop.hbase.log.HBaseMarkers; -045import org.apache.hadoop.hbase.procedure2.Procedure.LockState; -046import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; -047import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; -048import org.apache.hadoop.hbase.procedure2.util.StringUtils; -049import org.apache.hadoop.hbase.security.User; -050import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -051import org.apache.hadoop.hbase.util.IdLock; -052import org.apache.hadoop.hbase.util.NonceKey; -053import org.apache.hadoop.hbase.util.Threads; -054import org.apache.yetus.audience.InterfaceAudience; -055import org.slf4j.Logger; -056import org.slf4j.LoggerFactory; -057 -058import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -059import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; -060 -061import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; -062 -063/** -064 * Thread Pool that executes the submitted procedures. -065 * The executor has a ProcedureStore associated. -066 * Each operation is logged and on restart the pending procedures are resumed. -067 * -068 * Unless the Procedure code throws an error (e.g. invalid user input) -069 * the procedure will complete (at some point in time), On restart the pending -070 * procedures are resumed and the once failed will be rolledback. -071 * -072 * The user can add procedures to the executor via submitProcedure(proc) -073 * check for the finished state via isFinished(procId) -074 * and get the result via getResult(procId) -075 */ -076@InterfaceAudience.Private -077public class ProcedureExecutorTEnvironment { -078 private static final Logger LOG = LoggerFactory.getLogger(ProcedureExecutor.class); -079 -080 public static final String CHECK_OWNER_SET_CONF_KEY = "hbase.procedure.check.owner.set"; -081 private static final boolean DEFAULT_CHECK_OWNER_SET = false; -082 -083 public static final String WORKER_KEEP_ALIVE_TIME_CONF_KEY = -084 "hbase.procedure.worker.keep.alive.time.msec"; -085 private static final long DEFAULT_WORKER_KEEP_ALIVE_TIME = TimeUnit.MINUTES.toMillis(1); -086 -087 /** -088 * {@link #testing} is non-null when ProcedureExecutor is being tested. Tests will try to -089 * break PE having it fail at various junctures. When non-null, testing is set to an instance of -090 * the below internal {@link Testing} class with flags set for the particular test. -091 */ -092 Testing testing = null; -093 -094 /** -095 * Class with parameters describing how to fail/die when in testing-context. -096 */ -097 public static class Testing { -098protected boolean killIfHasParent = true; -099protected boolean killIfSuspended = false; -100 -101/** -102 * Kill the PE BEFORE we store state to the WAL. Good for figuring out if a Procedure is -103 * persisting all the state it needs to recover after a crash. -104 */ -105protected boolean killBeforeStoreUpdate = false; -106protected boolean toggleKillBeforeStoreUpdate = false; -107 -108/** -109 * Set when we want to fail AFTER state has been stored into the WAL. Rarely used. HBASE-20978 -110 * is about a case where memory-state was being set after store to WAL where a crash could -111 * cause us to get stuck. This flag allows killing at what was a vulnerable time. -112 */ -113protected boolean killAfterStoreUpdate = false; -114protected boolean toggleKillAfterStoreUpdate = false; -115 -116protected boolean shouldKillBeforeStoreUpdate() { -117 final boolean kill = this.killBeforeStoreUpdate; -118 if (this.toggleKillBeforeStoreUpdate) { -119this.killBeforeStoreUpdate = !kill; -120LOG.warn("Toggle KILL before store update to: " + this.killBeforeStoreUpdate); -121 } -122 return kill; -123} -124 -125protected boolean shouldKillBeforeStoreUpdate(boolean isSuspended, boolean hasParent) { -126 if (isSuspended
[04/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html b/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html index 0ba0d47..3a7cd7f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityLabelFilter.html @@ -31,77 +31,77 @@ 023 024import org.apache.yetus.audience.InterfaceAudience; 025import org.apache.hadoop.hbase.Cell; -026import org.apache.hadoop.hbase.CellUtil; -027import org.apache.hadoop.hbase.PrivateCellUtil; -028import org.apache.hadoop.hbase.filter.FilterBase; -029import org.apache.hadoop.hbase.util.ByteRange; -030import org.apache.hadoop.hbase.util.SimpleMutableByteRange; -031 -032/** -033 * This Filter checks the visibility expression with each KV against visibility labels associated -034 * with the scan. Based on the check the KV is included in the scan result or gets filtered out. -035 */ -036@InterfaceAudience.Private -037class VisibilityLabelFilter extends FilterBase { -038 -039 private final VisibilityExpEvaluator expEvaluator; -040 private final MapByteRange, Integer cfVsMaxVersions; -041 private final ByteRange curFamily; -042 private final ByteRange curQualifier; -043 private int curFamilyMaxVersions; -044 private int curQualMetVersions; -045 -046 public VisibilityLabelFilter(VisibilityExpEvaluator expEvaluator, -047 MapByteRange, Integer cfVsMaxVersions) { -048this.expEvaluator = expEvaluator; -049this.cfVsMaxVersions = cfVsMaxVersions; -050this.curFamily = new SimpleMutableByteRange(); -051this.curQualifier = new SimpleMutableByteRange(); -052 } -053 -054 @Override -055 public boolean filterRowKey(Cell cell) throws IOException { -056// Impl in FilterBase might do unnecessary copy for Off heap backed Cells. -057return false; -058 } -059 -060 @Override -061 public ReturnCode filterCell(final Cell cell) throws IOException { -062if (curFamily.getBytes() == null -063|| !(PrivateCellUtil.matchingFamily(cell, curFamily.getBytes(), curFamily.getOffset(), -064curFamily.getLength( { -065 curFamily.set(cell.getFamilyArray(), cell.getFamilyOffset(), cell.getFamilyLength()); -066 // For this family, all the columns can have max of curFamilyMaxVersions versions. No need to -067 // consider the older versions for visibility label check. -068 // Ideally this should have been done at a lower layer by HBase (?) -069 curFamilyMaxVersions = cfVsMaxVersions.get(curFamily); -070 // Family is changed. Just unset curQualifier. -071 curQualifier.unset(); -072} -073if (curQualifier.getBytes() == null || !(PrivateCellUtil.matchingQualifier(cell, -074 curQualifier.getBytes(), curQualifier.getOffset(), curQualifier.getLength( { -075 curQualifier.set(cell.getQualifierArray(), cell.getQualifierOffset(), -076 cell.getQualifierLength()); -077 curQualMetVersions = 0; -078} -079curQualMetVersions++; -080if (curQualMetVersions curFamilyMaxVersions) { -081 return ReturnCode.SKIP; -082} -083 -084return this.expEvaluator.evaluate(cell) ? ReturnCode.INCLUDE : ReturnCode.SKIP; -085 } -086 -087 @Override -088 public void reset() throws IOException { -089this.curFamily.unset(); -090this.curQualifier.unset(); -091this.curFamilyMaxVersions = 0; -092this.curQualMetVersions = 0; -093 } +026import org.apache.hadoop.hbase.PrivateCellUtil; +027import org.apache.hadoop.hbase.filter.FilterBase; +028import org.apache.hadoop.hbase.util.ByteRange; +029import org.apache.hadoop.hbase.util.SimpleMutableByteRange; +030 +031/** +032 * This Filter checks the visibility expression with each KV against visibility labels associated +033 * with the scan. Based on the check the KV is included in the scan result or gets filtered out. +034 */ +035@InterfaceAudience.Private +036class VisibilityLabelFilter extends FilterBase { +037 +038 private final VisibilityExpEvaluator expEvaluator; +039 private final MapByteRange, Integer cfVsMaxVersions; +040 private final ByteRange curFamily; +041 private final ByteRange curQualifier; +042 private int curFamilyMaxVersions; +043 private int curQualMetVersions; +044 +045 public VisibilityLabelFilter(VisibilityExpEvaluator expEvaluator, +046 MapByteRange, Integer cfVsMaxVersions) { +047this.expEvaluator = expEvaluator; +048this.cfVsMaxVersions = cfVsMaxVersions; +049this.curFamily = new SimpleMutableByteRange(); +050this.curQualifier = new SimpleMutableByteRange(); +051 } +052 +053 @Override +054 public boolean
[04/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureUtil.CompatStateSerializer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureUtil.CompatStateSerializer.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureUtil.CompatStateSerializer.html index 1a491f3..76e3c88 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureUtil.CompatStateSerializer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureUtil.CompatStateSerializer.html @@ -213,122 +213,130 @@ 205if (proc.hasLock()) { 206 builder.setLocked(true); 207} -208return builder.build(); -209 } -210 -211 /** -212 * Helper to convert the protobuf procedure. -213 * Used by ProcedureStore implementations. -214 * -215 * TODO: OPTIMIZATION: some of the field never change during the execution -216 * (e.g. className, procId, parentId, ...). -217 * We can split in 'data' and 'state', and the store -218 * may take advantage of it by storing the data only on insert(). -219 */ -220 public static Procedure convertToProcedure(final ProcedureProtos.Procedure proto) throws IOException { -221// Procedure from class name -222final Procedure proc = newProcedure(proto.getClassName()); -223 -224// set fields -225proc.setProcId(proto.getProcId()); -226proc.setState(proto.getState()); -227 proc.setSubmittedTime(proto.getSubmittedTime()); -228 proc.setLastUpdate(proto.getLastUpdate()); -229 -230if (proto.hasParentId()) { -231 proc.setParentProcId(proto.getParentId()); -232} +208 +209if (proc.isBypass()) { +210 builder.setBypass(true); +211} +212return builder.build(); +213 } +214 +215 /** +216 * Helper to convert the protobuf procedure. +217 * Used by ProcedureStore implementations. +218 * +219 * TODO: OPTIMIZATION: some of the field never change during the execution +220 * (e.g. className, procId, parentId, ...). +221 * We can split in 'data' and 'state', and the store +222 * may take advantage of it by storing the data only on insert(). +223 */ +224 public static Procedure convertToProcedure(final ProcedureProtos.Procedure proto) throws IOException { +225// Procedure from class name +226final Procedure proc = newProcedure(proto.getClassName()); +227 +228// set fields +229proc.setProcId(proto.getProcId()); +230proc.setState(proto.getState()); +231 proc.setSubmittedTime(proto.getSubmittedTime()); +232 proc.setLastUpdate(proto.getLastUpdate()); 233 -234if (proto.hasOwner()) { -235 proc.setOwner(proto.getOwner()); +234if (proto.hasParentId()) { +235 proc.setParentProcId(proto.getParentId()); 236} 237 -238if (proto.hasTimeout()) { -239 proc.setTimeout(proto.getTimeout()); +238if (proto.hasOwner()) { +239 proc.setOwner(proto.getOwner()); 240} 241 -242if (proto.getStackIdCount() 0) { -243 proc.setStackIndexes(proto.getStackIdList()); +242if (proto.hasTimeout()) { +243 proc.setTimeout(proto.getTimeout()); 244} 245 -246if (proto.hasException()) { -247 assert proc.getState() == ProcedureProtos.ProcedureState.FAILED || -248 proc.getState() == ProcedureProtos.ProcedureState.ROLLEDBACK : -249 "The procedure must be failed (waiting to rollback) or rolledback"; -250 proc.setFailure(RemoteProcedureException.fromProto(proto.getException())); -251} -252 -253if (proto.hasResult()) { -254 proc.setResult(proto.getResult().toByteArray()); +246if (proto.getStackIdCount() 0) { +247 proc.setStackIndexes(proto.getStackIdList()); +248} +249 +250if (proto.hasException()) { +251 assert proc.getState() == ProcedureProtos.ProcedureState.FAILED || +252 proc.getState() == ProcedureProtos.ProcedureState.ROLLEDBACK : +253 "The procedure must be failed (waiting to rollback) or rolledback"; +254 proc.setFailure(RemoteProcedureException.fromProto(proto.getException())); 255} 256 -257if (proto.getNonce() != HConstants.NO_NONCE) { -258 proc.setNonceKey(new NonceKey(proto.getNonceGroup(), proto.getNonce())); +257if (proto.hasResult()) { +258 proc.setResult(proto.getResult().toByteArray()); 259} 260 -261if (proto.getLocked()) { -262 proc.lockedWhenLoading(); +261if (proto.getNonce() != HConstants.NO_NONCE) { +262 proc.setNonceKey(new NonceKey(proto.getNonceGroup(), proto.getNonce())); 263} 264 -265ProcedureStateSerializer serializer = null; -266 -267if (proto.getStateMessageCount() 0) { -268 serializer = new
[04/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/testdevapidocs/org/apache/hadoop/hbase/client/package-summary.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/client/package-summary.html b/testdevapidocs/org/apache/hadoop/hbase/client/package-summary.html index 5b020ba..ccfe07f 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/client/package-summary.html +++ b/testdevapidocs/org/apache/hadoop/hbase/client/package-summary.html @@ -873,385 +873,391 @@ +TestHbck + +Class to test HBaseHbck. + + + TestHTableMultiplexer - + TestHTableMultiplexerFlushCache - + TestHTableMultiplexerViaMocks - + TestIllegalTableDescriptor - + TestImmutableHColumnDescriptor - + TestImmutableHRegionInfo Test ImmutableHRegionInfo - + TestImmutableHTableDescriptor - + TestIncrement - + TestIncrementFromClientSideWithCoprocessor Test all Increment client operations with a coprocessor that just implements the default flush/compact/scan policy. - + TestIncrementsFromClientSide Run Increment tests that use the HBase clients; HTable. - + TestInterfaceAlign - + TestIntraRowPagination Test scan/get offset and limit settings within one row through HRegion API. - + TestLeaseRenewal - + TestLimitedScanWithFilter With filter we may stop at a middle of row and think that we still have more cells for the current row but actually all the remaining cells will be filtered out by the filter. - + TestMalformedCellFromClient The purpose of this test is to ensure whether rs deals with the malformed cells correctly. - + TestMetaCache - + TestMetaCache.CallQueueTooBigExceptionInjector Throws CallQueueTooBigException for all gets. - + TestMetaCache.ExceptionInjector - + TestMetaCache.FakeRSRpcServices - + TestMetaCache.RegionServerWithFakeRpcServices - + TestMetaCache.RoundRobinExceptionInjector Rotates through the possible cache clearing and non-cache clearing exceptions for requests. - + TestMetaWithReplicas Tests the scenarios where replicas are enabled for the meta table - + TestMetricsConnection - + TestMobCloneSnapshotFromClient Test clone snapshots from the client - + TestMobCloneSnapshotFromClient.DelayFlushCoprocessor This coprocessor is used to delay the flush. - + TestMobRestoreSnapshotFromClient Test restore snapshots from the client - + TestMobSnapshotCloneIndependence Test to verify that the cloned table is independent of the table from which it was cloned - + TestMobSnapshotFromClient Test create/using/deleting snapshots from the client - + TestMultiParallel - + TestMultiParallel.MyMasterObserver - + TestMultipleTimestamps Run tests related to TimestampsFilter using HBase client APIs. - + TestMultiRespectsLimits This test sets the multi size WAY low and then checks to make sure that gets will still make progress. - + TestMutation - + TestMvccConsistentScanner - + TestOperation Run tests that use the functionality of the Operation superclass for Puts, Gets, Deletes, Scans, and MultiPuts. - + TestProcedureFuture - + TestProcedureFuture.TestFuture - + TestPutDeleteEtcCellIteration Test that I can Iterate Client Actions that hold Cells (Get does not have Cells). - + TestPutDotHas - + TestPutWithDelete - + TestPutWriteToWal - + TestQuotasShell - + TestRawAsyncScanCursor - + TestRawAsyncTableLimitedScanWithFilter With filter we may stop at a middle of row and think that we still have more cells for the current row but actually all the remaining cells will be filtered out by the filter. - + TestRawAsyncTablePartialScan - + TestRawAsyncTableScan - + TestRegionInfoDisplay - + TestReplicasClient Tests for region replicas. - + TestReplicasClient.SlowMeCopro This copro is used to synchronize the tests. - + TestReplicationShell - + TestReplicaWithCluster - + TestReplicaWithCluster.RegionServerHostingPrimayMetaRegionSlowOrStopCopro This copro is used to slow down the primary meta region scan a bit - + TestReplicaWithCluster.RegionServerStoppedCopro This copro is used to simulate region server down exception for Get and Scan - + TestReplicaWithCluster.SlowMeCopro This copro is used to synchronize the tests. - + TestRestoreSnapshotFromClient Test restore snapshots from the client - + TestRestoreSnapshotFromClientWithRegionReplicas - + TestResult - + TestResultFromCoprocessor - + TestResultFromCoprocessor.MyObserver - + TestResultScannerCursor - + TestResultSizeEstimation - + TestRetriesExhaustedWithDetailsException - + TestReversedScannerCallable - + TestRowComparator - + TestRpcControllerFactory - +
[04/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/devapidocs/src-html/org/apache/hadoop/hbase/filter/ColumnValueFilter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/filter/ColumnValueFilter.html b/devapidocs/src-html/org/apache/hadoop/hbase/filter/ColumnValueFilter.html index ee5595f..961079a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/ColumnValueFilter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/ColumnValueFilter.html @@ -29,224 +29,240 @@ 021 022import java.io.IOException; 023import java.util.ArrayList; -024 -025import org.apache.hadoop.hbase.Cell; -026import org.apache.hadoop.hbase.CellUtil; -027import org.apache.hadoop.hbase.CompareOperator; -028import org.apache.hadoop.hbase.PrivateCellUtil; -029import org.apache.hadoop.hbase.exceptions.DeserializationException; -030import org.apache.hadoop.hbase.util.Bytes; -031import org.apache.yetus.audience.InterfaceAudience; -032 -033import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; -034import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException; -035import org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations; -036 -037import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -038import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos; -039import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; -040 -041/** -042 * Different from {@link SingleColumnValueFilter} which returns an bentire/b row -043 * when specified condition is matched, {@link ColumnValueFilter} return the matched cell only. -044 * p -045 * This filter is used to filter cells based on column and value. -046 * It takes a {@link org.apache.hadoop.hbase.CompareOperator} operator (, =, =, !=, , =), and -047 * and a {@link ByteArrayComparable} comparator. -048 */ -049@InterfaceAudience.Public -050public class ColumnValueFilter extends FilterBase { -051 private final byte[] family; -052 private final byte[] qualifier; -053 private final CompareOperator op; -054 private final ByteArrayComparable comparator; -055 -056 // This flag is used to speed up seeking cells when matched column is found, such that following -057 // columns in the same row can be skipped faster by NEXT_ROW instead of NEXT_COL. -058 private boolean columnFound = false; -059 -060 public ColumnValueFilter(final byte[] family, final byte[] qualifier, -061 final CompareOperator op, final byte[] value) { -062this(family, qualifier, op, new BinaryComparator(value)); -063 } -064 -065 public ColumnValueFilter(final byte[] family, final byte[] qualifier, -066 final CompareOperator op, -067 final ByteArrayComparable comparator) { -068this.family = Preconditions.checkNotNull(family, "family should not be null."); -069this.qualifier = qualifier == null ? new byte[0] : qualifier; -070this.op = Preconditions.checkNotNull(op, "CompareOperator should not be null"); -071this.comparator = Preconditions.checkNotNull(comparator, "Comparator should not be null"); -072 } -073 -074 /** -075 * @return operator -076 */ -077 public CompareOperator getCompareOperator() { -078return op; -079 } -080 -081 /** -082 * @return the comparator -083 */ -084 public ByteArrayComparable getComparator() { -085return comparator; -086 } -087 -088 /** -089 * @return the column family -090 */ -091 public byte[] getFamily() { -092return family; -093 } -094 -095 /** -096 * @return the qualifier -097 */ -098 public byte[] getQualifier() { -099return qualifier; -100 } -101 -102 @Override -103 public void reset() throws IOException { -104columnFound = false; -105 } -106 -107 @Override -108 public boolean filterRowKey(Cell cell) throws IOException { -109return false; -110 } -111 -112 @Override -113 public ReturnCode filterCell(Cell c) throws IOException { -114// 1. Check column match -115if (!CellUtil.matchingColumn(c, this.family, this.qualifier)) { -116 return columnFound ? ReturnCode.NEXT_ROW : ReturnCode.NEXT_COL; -117} -118// Column found -119columnFound = true; -120// 2. Check value match: -121// True means filter out, just skip this cell, else include it. -122return compareValue(getCompareOperator(), getComparator(), c) ? -123 ReturnCode.SKIP : ReturnCode.INCLUDE; -124 } -125 -126 /** -127 * This method is used to determine a cell should be included or filtered out. -128 * @param op one of operators {@link CompareOperator} -129 * @param comparator comparator used to compare cells. -130 * @param cell cell to be compared. -131 * @return true means cell should be filtered out, included otherwise. -132 */ -133 private boolean compareValue(final CompareOperator op, final
[04/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html index edfb4bf..8b112b5 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html +++ b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestHRegion.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":10,"i2":10,"i3":10,"i4":9,"i5":9,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":9,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":9,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109": 10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":9,"i129":10}; +var methods = {"i0":9,"i1":10,"i2":10,"i3":10,"i4":9,"i5":9,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":9,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":9,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109": 10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":9,"i130":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -644,235 +644,241 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html void +testFlushAndMemstoreSizeCounting() +A test case of HBASE-21041 + + + +void testFlushCacheWhileScanning() Flushes the cache in a thread while scanning. - + void testFlushedFileWithNoTags() - + void testFlushMarkers() - + void testFlushMarkersWALFail() - + void testFlushResult() Test that we get the expected flush results back - + void testFlushSizeAccounting() Test we do not lose data if we fail a flush and then close. - + void testGet_Basic() - + void testGet_Empty() - + void testGet_FamilyChecker() - + void testgetHDFSBlocksDistribution() - + void testGetScanner_WithNoFamilies() - + void testGetScanner_WithNotOkFamilies() - + void testGetScanner_WithOkFamilies() - + void testGetScanner_WithRegionClosed() This method tests https://issues.apache.org/jira/browse/HBASE-2516. - + void testGetWhileRegionClose() - + void testGetWithFilter() - + void testHolesInMeta() - + void testIncrementTimestampsAreMonotonic() - + void testIncrWithReadOnlyTable() - + void testIndexesScanWithOneDeletedRow() - + void testLongQualifier() Write an HFile block full with Cells whose qualifier that are identical between 0 and Short.MAX_VALUE. - + void testMemstoreSizeAccountingWithFailedPostBatchMutate() - + void testMemstoreSnapshotSize() - + void testMutateRow_WriteRequestCount() - + void testOpenRegionWrittenToWAL() - + void testParallelAppendWithMemStoreFlush() Test case to check append function with memstore flushing - + void testParallelIncrementWithMemStoreFlush() Test case to check increment function with memstore flushing - + void testPutWithLatestTS()
[04/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html index 81f5178..7df71bd 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html @@ -108,3669 +108,3727 @@ 100import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; 101import org.apache.hadoop.hbase.log.HBaseMarkers; 102import org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode; -103import org.apache.hadoop.hbase.master.assignment.AssignmentManager; -104import org.apache.hadoop.hbase.master.assignment.MergeTableRegionsProcedure; -105import org.apache.hadoop.hbase.master.assignment.RegionStates; -106import org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode; -107import org.apache.hadoop.hbase.master.balancer.BalancerChore; -108import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer; -109import org.apache.hadoop.hbase.master.balancer.ClusterStatusChore; -110import org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory; -111import org.apache.hadoop.hbase.master.cleaner.CleanerChore; -112import org.apache.hadoop.hbase.master.cleaner.HFileCleaner; -113import org.apache.hadoop.hbase.master.cleaner.LogCleaner; -114import org.apache.hadoop.hbase.master.cleaner.ReplicationBarrierCleaner; -115import org.apache.hadoop.hbase.master.locking.LockManager; -116import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -117import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType; -118import org.apache.hadoop.hbase.master.normalizer.RegionNormalizer; -119import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore; -120import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory; -121import org.apache.hadoop.hbase.master.procedure.CreateTableProcedure; -122import org.apache.hadoop.hbase.master.procedure.DeleteNamespaceProcedure; -123import org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure; -124import org.apache.hadoop.hbase.master.procedure.DisableTableProcedure; -125import org.apache.hadoop.hbase.master.procedure.EnableTableProcedure; -126import org.apache.hadoop.hbase.master.procedure.InitMetaProcedure; -127import org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants; -128import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; -129import org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler; -130import org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil; -131import org.apache.hadoop.hbase.master.procedure.ModifyTableProcedure; -132import org.apache.hadoop.hbase.master.procedure.ProcedurePrepareLatch; -133import org.apache.hadoop.hbase.master.procedure.ServerCrashProcedure; -134import org.apache.hadoop.hbase.master.procedure.TruncateTableProcedure; -135import org.apache.hadoop.hbase.master.replication.AbstractPeerProcedure; -136import org.apache.hadoop.hbase.master.replication.AddPeerProcedure; -137import org.apache.hadoop.hbase.master.replication.DisablePeerProcedure; -138import org.apache.hadoop.hbase.master.replication.EnablePeerProcedure; -139import org.apache.hadoop.hbase.master.replication.RemovePeerProcedure; -140import org.apache.hadoop.hbase.master.replication.ReplicationPeerManager; -141import org.apache.hadoop.hbase.master.replication.SyncReplicationReplayWALManager; -142import org.apache.hadoop.hbase.master.replication.TransitPeerSyncReplicationStateProcedure; -143import org.apache.hadoop.hbase.master.replication.UpdatePeerConfigProcedure; -144import org.apache.hadoop.hbase.master.snapshot.SnapshotManager; -145import org.apache.hadoop.hbase.master.zksyncer.MasterAddressSyncer; -146import org.apache.hadoop.hbase.master.zksyncer.MetaLocationSyncer; -147import org.apache.hadoop.hbase.mob.MobConstants; -148import org.apache.hadoop.hbase.monitoring.MemoryBoundedLogMessageBuffer; -149import org.apache.hadoop.hbase.monitoring.MonitoredTask; -150import org.apache.hadoop.hbase.monitoring.TaskMonitor; -151import org.apache.hadoop.hbase.procedure.MasterProcedureManagerHost; -152import org.apache.hadoop.hbase.procedure.flush.MasterFlushTableProcedureManager; -153import org.apache.hadoop.hbase.procedure2.LockedResource; -154import org.apache.hadoop.hbase.procedure2.Procedure; -155import org.apache.hadoop.hbase.procedure2.ProcedureEvent; -156import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; -157import org.apache.hadoop.hbase.procedure2.RemoteProcedureDispatcher.RemoteProcedure; -158import org.apache.hadoop.hbase.procedure2.RemoteProcedureException;
[04/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.html b/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.html index 0b475e4..6b1d637 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.html @@ -52,126 +52,130 @@ 044import org.apache.zookeeper.server.ZooKeeperServerMain; 045import org.apache.zookeeper.server.quorum.QuorumPeerConfig; 046import org.apache.zookeeper.server.quorum.QuorumPeerMain; -047 -048/** -049 * HBase's version of ZooKeeper's QuorumPeer. When HBase is set to manage -050 * ZooKeeper, this class is used to start up QuorumPeer instances. By doing -051 * things in here rather than directly calling to ZooKeeper, we have more -052 * control over the process. This class uses {@link ZKConfig} to get settings -053 * from the hbase-site.xml file. -054 */ -055@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -056@InterfaceStability.Evolving -057public final class HQuorumPeer { -058 private HQuorumPeer() { -059 } -060 -061 /** -062 * Parse ZooKeeper configuration from HBase XML config and run a QuorumPeer. -063 * @param args String[] of command line arguments. Not used. -064 */ -065 public static void main(String[] args) { -066Configuration conf = HBaseConfiguration.create(); -067try { -068 Properties zkProperties = ZKConfig.makeZKProps(conf); -069 writeMyID(zkProperties); -070 QuorumPeerConfig zkConfig = new QuorumPeerConfig(); -071 zkConfig.parseProperties(zkProperties); -072 -073 // login the zookeeper server principal (if using security) -074 ZKUtil.loginServer(conf, HConstants.ZK_SERVER_KEYTAB_FILE, -075 HConstants.ZK_SERVER_KERBEROS_PRINCIPAL, -076 zkConfig.getClientPortAddress().getHostName()); -077 -078 runZKServer(zkConfig); -079} catch (Exception e) { -080 e.printStackTrace(); -081 System.exit(-1); -082} -083 } -084 -085 private static void runZKServer(QuorumPeerConfig zkConfig) -086 throws UnknownHostException, IOException { -087if (zkConfig.isDistributed()) { -088 QuorumPeerMain qp = new QuorumPeerMain(); -089 qp.runFromConfig(zkConfig); -090} else { -091 ZooKeeperServerMain zk = new ZooKeeperServerMain(); -092 ServerConfig serverConfig = new ServerConfig(); -093 serverConfig.readFrom(zkConfig); -094 zk.runFromConfig(serverConfig); -095} -096 } -097 -098 private static boolean addressIsLocalHost(String address) { -099return address.equals("localhost") || address.equals("127.0.0.1"); +047import org.slf4j.Logger; +048import org.slf4j.LoggerFactory; +049 +050/** +051 * HBase's version of ZooKeeper's QuorumPeer. When HBase is set to manage +052 * ZooKeeper, this class is used to start up QuorumPeer instances. By doing +053 * things in here rather than directly calling to ZooKeeper, we have more +054 * control over the process. This class uses {@link ZKConfig} to get settings +055 * from the hbase-site.xml file. +056 */ +057@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) +058@InterfaceStability.Evolving +059public final class HQuorumPeer { +060 private static final Logger LOG = LoggerFactory.getLogger(HQuorumPeer.class); +061 +062 private HQuorumPeer() { +063 } +064 +065 /** +066 * Parse ZooKeeper configuration from HBase XML config and run a QuorumPeer. +067 * @param args String[] of command line arguments. Not used. +068 */ +069 public static void main(String[] args) { +070Configuration conf = HBaseConfiguration.create(); +071try { +072 Properties zkProperties = ZKConfig.makeZKProps(conf); +073 writeMyID(zkProperties); +074 QuorumPeerConfig zkConfig = new QuorumPeerConfig(); +075 zkConfig.parseProperties(zkProperties); +076 +077 // login the zookeeper server principal (if using security) +078 ZKUtil.loginServer(conf, HConstants.ZK_SERVER_KEYTAB_FILE, +079 HConstants.ZK_SERVER_KERBEROS_PRINCIPAL, +080 zkConfig.getClientPortAddress().getHostName()); +081 +082 runZKServer(zkConfig); +083} catch (Exception e) { +084 LOG.error("Failed to start ZKServer", e); +085 System.exit(-1); +086} +087 } +088 +089 private static void runZKServer(QuorumPeerConfig zkConfig) +090 throws UnknownHostException, IOException { +091if (zkConfig.isDistributed()) { +092 QuorumPeerMain qp = new QuorumPeerMain(); +093 qp.runFromConfig(zkConfig); +094} else { +095 ZooKeeperServerMain zk = new ZooKeeperServerMain(); +096 ServerConfig serverConfig = new ServerConfig(); +097 serverConfig.readFrom(zkConfig); +098
[04/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/testdevapidocs/allclasses-frame.html -- diff --git a/testdevapidocs/allclasses-frame.html b/testdevapidocs/allclasses-frame.html index 1b24e50..b9a38d0 100644 --- a/testdevapidocs/allclasses-frame.html +++ b/testdevapidocs/allclasses-frame.html @@ -490,6 +490,7 @@ RESTApiClusterManager.Service RestartActionBaseAction RestartActiveMasterAction +RestartActiveNameNodeAction RestartMetaTest RestartRandomDataNodeAction RestartRandomRsAction http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/testdevapidocs/allclasses-noframe.html -- diff --git a/testdevapidocs/allclasses-noframe.html b/testdevapidocs/allclasses-noframe.html index 7b96d5d..8092c7f 100644 --- a/testdevapidocs/allclasses-noframe.html +++ b/testdevapidocs/allclasses-noframe.html @@ -490,6 +490,7 @@ RESTApiClusterManager.Service RestartActionBaseAction RestartActiveMasterAction +RestartActiveNameNodeAction RestartMetaTest RestartRandomDataNodeAction RestartRandomRsAction http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/testdevapidocs/constant-values.html -- diff --git a/testdevapidocs/constant-values.html b/testdevapidocs/constant-values.html index 92fdfe5..6491347 100644 --- a/testdevapidocs/constant-values.html +++ b/testdevapidocs/constant-values.html @@ -2220,6 +2220,20 @@ "hbase.chaosmonkey.action.killmastertimeout" + + +protectedstaticfinallong +KILL_NAMENODE_TIMEOUT_DEFAULT +6L + + + + +publicstaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +KILL_NAMENODE_TIMEOUT_KEY +"hbase.chaosmonkey.action.killnamenodetimeout" + + protectedstaticfinallong @@ -2276,6 +2290,20 @@ "hbase.chaosmonkey.action.startmastertimeout" + + +protectedstaticfinallong +START_NAMENODE_TIMEOUT_DEFAULT +6L + + + + +publicstaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +START_NAMENODE_TIMEOUT_KEY +"hbase.chaosmonkey.action.startnamenodetimeout" + + protectedstaticfinallong @@ -2327,6 +2355,39 @@ +org.apache.hadoop.hbase.chaos.actions.RestartActiveNameNodeAction + +Modifier and Type +Constant Field +Value + + + + + +privatestaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +ACTIVE_NN_LOCK_NAME +"ActiveStandbyElectorLock" + + + + +privatestaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +ZK_PARENT_ZNODE_DEFAULT +"/hadoop-ha" + + + + +privatestaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +ZK_PARENT_ZNODE_KEY +"ha.zookeeper.parent-znode" + + + + + + org.apache.hadoop.hbase.chaos.actions.SplitAllRegionOfTableAction Modifier and Type http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/testdevapidocs/index-all.html -- diff --git a/testdevapidocs/index-all.html b/testdevapidocs/index-all.html index 8c8b1c3..41763ca 100644 --- a/testdevapidocs/index-all.html +++ b/testdevapidocs/index-all.html @@ -498,6 +498,8 @@ activateFailure - Static variable in class org.apache.hadoop.hbase.regionserver.wal.InstrumentedLogWriter +ACTIVE_NN_LOCK_NAME - Static variable in class org.apache.hadoop.hbase.chaos.actions.RestartActiveNameNodeAction + activeMasterManager - Variable in class org.apache.hadoop.hbase.master.TestActiveMasterManager.DummyMaster ACTOR_PATTERN - Static variable in class org.apache.hadoop.hbase.mapred.TestTableMapReduceUtil @@ -22257,6 +22259,10 @@ KILL_MASTER_TIMEOUT_KEY - Static variable in class org.apache.hadoop.hbase.chaos.actions.Action +KILL_NAMENODE_TIMEOUT_DEFAULT - Static variable in class org.apache.hadoop.hbase.chaos.actions.Action + +KILL_NAMENODE_TIMEOUT_KEY - Static variable in class org.apache.hadoop.hbase.chaos.actions.Action + KILL_RS_TIMEOUT_DEFAULT - Static variable in class org.apache.hadoop.hbase.chaos.actions.Action KILL_RS_TIMEOUT_KEY - Static variable in class org.apache.hadoop.hbase.chaos.actions.Action @@ -22319,6 +22325,19 @@ killMetaRs - Variable in class org.apache.hadoop.hbase.chaos.factories.UnbalanceMonkeyFactory +killNameNode(ServerName) - Method in class org.apache.hadoop.hbase.chaos.actions.Action + +killNameNode(ServerName) - Method in class org.apache.hadoop.hbase.DistributedHBaseCluster + +killNameNode(ServerName) - Method in class org.apache.hadoop.hbase.HBaseCluster + +Kills the namenode process if this is a distributed
[04/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html index dd41a30..624224f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html @@ -90,768 +90,758 @@ 082extends AbstractStateMachineTableProcedureMergeTableRegionsState { 083 private static final Logger LOG = LoggerFactory.getLogger(MergeTableRegionsProcedure.class); 084 private Boolean traceEnabled; -085 private volatile boolean lock = false; -086 private ServerName regionLocation; -087 private RegionInfo[] regionsToMerge; -088 private RegionInfo mergedRegion; -089 private boolean forcible; -090 -091 public MergeTableRegionsProcedure() { -092// Required by the Procedure framework to create the procedure on replay -093 } -094 -095 public MergeTableRegionsProcedure(final MasterProcedureEnv env, -096 final RegionInfo regionToMergeA, final RegionInfo regionToMergeB) throws IOException { -097this(env, regionToMergeA, regionToMergeB, false); -098 } -099 -100 public MergeTableRegionsProcedure(final MasterProcedureEnv env, -101 final RegionInfo regionToMergeA, final RegionInfo regionToMergeB, -102 final boolean forcible) throws IOException { -103this(env, new RegionInfo[] {regionToMergeA, regionToMergeB}, forcible); -104 } -105 -106 public MergeTableRegionsProcedure(final MasterProcedureEnv env, -107 final RegionInfo[] regionsToMerge, final boolean forcible) -108 throws IOException { -109super(env); -110 -111// Check daughter regions and make sure that we have valid daughter regions -112// before doing the real work. This check calls the super method #checkOnline also. -113checkRegionsToMerge(env, regionsToMerge, forcible); -114 -115// WARN: make sure there is no parent region of the two merging regions in -116// hbase:meta If exists, fixing up daughters would cause daughter regions(we -117// have merged one) online again when we restart master, so we should clear -118// the parent region to prevent the above case -119// Since HBASE-7721, we don't need fix up daughters any more. so here do nothing -120this.regionsToMerge = regionsToMerge; -121this.mergedRegion = createMergedRegionInfo(regionsToMerge); -122preflightChecks(env, true); -123this.forcible = forcible; -124 } -125 -126 private static void checkRegionsToMerge(MasterProcedureEnv env, final RegionInfo[] regionsToMerge, -127 final boolean forcible) throws MergeRegionException { -128// For now, we only merge 2 regions. -129// It could be extended to more than 2 regions in the future. -130if (regionsToMerge == null || regionsToMerge.length != 2) { -131 throw new MergeRegionException("Expected to merge 2 regions, got: " + -132 Arrays.toString(regionsToMerge)); -133} -134 -135checkRegionsToMerge(env, regionsToMerge[0], regionsToMerge[1], forcible); -136 } -137 -138 /** -139 * One time checks. -140 */ -141 private static void checkRegionsToMerge(MasterProcedureEnv env, final RegionInfo regionToMergeA, -142 final RegionInfo regionToMergeB, final boolean forcible) throws MergeRegionException { -143if (!regionToMergeA.getTable().equals(regionToMergeB.getTable())) { -144 throw new MergeRegionException("Can't merge regions from two different tables: " + -145regionToMergeA + ", " + regionToMergeB); -146} -147 -148if (regionToMergeA.getReplicaId() != RegionInfo.DEFAULT_REPLICA_ID || -149regionToMergeB.getReplicaId() != RegionInfo.DEFAULT_REPLICA_ID) { -150 throw new MergeRegionException("Can't merge non-default replicas"); -151} -152 -153try { -154 checkOnline(env, regionToMergeA); -155 checkOnline(env, regionToMergeB); -156} catch (DoNotRetryRegionException dnrre) { -157 throw new MergeRegionException(dnrre); -158} -159 -160if (!RegionInfo.areAdjacent(regionToMergeA, regionToMergeB)) { -161 String msg = "Unable to merge non-adjacent regions " + regionToMergeA.getShortNameToLog() + -162 ", " + regionToMergeB.getShortNameToLog() + " where forcible = " + forcible; -163 LOG.warn(msg); -164 if (!forcible) { -165throw new MergeRegionException(msg); -166 } -167} -168 } +085 private ServerName regionLocation; +086 private RegionInfo[] regionsToMerge; +087 private RegionInfo mergedRegion; +088 private boolean forcible; +089 +090 public MergeTableRegionsProcedure() { +091//
[04/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureDeque.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureDeque.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureDeque.html index 7766509..1b983c2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureDeque.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureDeque.html @@ -39,7 +39,11 @@ 031 */ 032@InterfaceAudience.Private 033public class ProcedureDeque extends ArrayDequeProcedure { -034} +034 public ProcedureDeque() { +035// Default is 16 for a list that is rarely used; elements will resize if too small. +036super(2); +037 } +038}
[04/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/class-use/HColumnDescriptor.html -- diff --git a/apidocs/org/apache/hadoop/hbase/class-use/HColumnDescriptor.html b/apidocs/org/apache/hadoop/hbase/class-use/HColumnDescriptor.html index 03beb7e..9b34e4d 100644 --- a/apidocs/org/apache/hadoop/hbase/class-use/HColumnDescriptor.html +++ b/apidocs/org/apache/hadoop/hbase/class-use/HColumnDescriptor.html @@ -1,10 +1,10 @@ http://www.w3.org/TR/html4/loose.dtd;> - + -ç±» org.apache.hadoop.hbase.HColumnDescriptorçä½¿ç¨ (Apache HBase 3.0.0-SNAPSHOT API) +Uses of Class org.apache.hadoop.hbase.HColumnDescriptor (Apache HBase 3.0.0-SNAPSHOT API) @@ -12,7 +12,7 @@ -æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã +JavaScript is disabled on your browser. -è·³è¿å¯¼èªé¾æ¥ +Skip navigation links - -æ¦è§ -ç¨åºå -ç±» -ä½¿ç¨ -æ -å·²è¿æ¶ -ç´¢å¼ -å¸®å© + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help -ä¸ä¸ä¸ª -ä¸ä¸ä¸ª +Prev +Next -æ¡æ¶ -æ æ¡æ¶ +Frames +NoFrames -ææç±» +AllClasses-ç±»ç使ç¨
+
org.apache.hadoop.hbase.HColumnDescriptorUses of Class
org.apache.hadoop.hbase.HColumnDescriptor
- -